Apps tagged with 'llm-inference'

All apps in Apps tagged with 'llm-inference' category. Use the filters below to narrow down your search. 
Copy a direct link to this comment to your clipboard
  1. Fireworks AI icon
     2 likes

    Build and deploy generative AI on the fastest and most efficient inference engine, fine-tuning and switching between models without extra costs.

    Cost / License

    • Paid
    • Proprietary

    Platforms

    • Online
    Fireworks AI screenshot 1
    Fireworks AI screenshot 1
    Fireworks AI screenshot 2
    +1
    Fireworks AI screenshot 3
    23 alternatives
  2. Learn how to add AI with local models and APIs to Windows apps. Discover AI scenarios and models such as Phi, Mistral, Stable Diffusion, Whisper, and many more to delight your users. The AI Dev Gallery is an open-source app designed to help Windows developers integrate AI...

    Cost / License

    • Free
    • Open Source (MIT)

    Platforms

    • Windows
    AI Dev Gallery screenshot 1
    AI Dev Gallery screenshot 1
    AI Dev Gallery screenshot 2
    +1
    AI Dev Gallery screenshot 3
    24 alternatives
  3. Lemonade helps users discover and run local AI apps by serving optimized LLMs right from their own GPUs and NPUs.

    Cost / License

    Platforms

    • Windows
    • Linux
    • Docker
    • Snapcraft
    • iPhone
    • iPad
    • Self-Hosted
    • Python
    Lemonade Server screenshot 1
    Lemonade Server screenshot 1
    Lemonade Server screenshot 2
    +1
    Lemonade Server screenshot 3
    16 alternatives
  4. AI Playground icon
     2 likes

    This application provides a full suite of generative AI features for chat, code assistance, document search, image analysis, image and video generation. All features run offline and are powered by your PC’s Intel® Core™ Ultra with built-in Intel Arc GPU or Intel Arc™ dGPU...

    Cost / License

    • Free
    • Open Source (MIT)

    Platforms

    • Windows
    AI Playground screenshot 1
    AI Playground screenshot 1
    23 alternatives
  5. RamaLama icon
     1 like

    RamaLama is an open-source developer tool that simplifies the local serving of AI models from any source and facilitates their use for inference in production, all through the familiar language of containers.

    Cost / License

    • Free
    • Open Source (MIT)

    Platforms

    • Linux
    • Mac
    • Python
    54 alternatives
  6. OpenRouter Runner is a monolith inference engine, built with Modal. It serves as a robust solution for the deployment of tons of open source models that are hosted in a fallback capacity on openrouter.ai.

    Cost / License

    • Freemium
    • Open Source (MIT)

    Platforms

    • Self-Hosted
  7. Corpus2GPT icon
     1 like

    Corpus2GPT: A project enabling users to train their own GPT models on diverse datasets, including local languages and various corpus types, using Keras and compatible with TensorFlow, PyTorch, or JAX backends for subsequent storage or sharing.

    Cost / License

    Platforms

    • Self-Hosted
    Training GPT from scratch in an artificial dummy data in 50 lines of code!
    5 alternatives
  8. Arch icon
     Like

    The proxy server for AI-native apps. Arch handles the pesky low-level work in building agents like clariyfing vague user input, routing prompts to the right agents and unifying access to any LLM - all without locking you into a framework.

    Cost / License

    Platforms

    • Self-Hosted
    • Docker
    Arch screenshot 1
  9. The DeepSeek API provides developers with direct access to DeepSeek’s advanced AI models, enabling them to run text, code, and multimodal tasks through simple endpoints for seamless integration into applications.

    Cost / License

    • Paid
    • Proprietary

    Platforms

    • Online
    DeepSeek Platform screenshot 1
    DeepSeek Platform screenshot 1
    DeepSeek Platform screenshot 2
    +1
    DeepSeek Platform screenshot 3
    12 alternatives
  10. GPUniq icon
     Like

    GPUniq is a stable GPU compute platform built for AI teams and developers who need reliable performance without paying traditional cloud prices. It provides on-demand access to powerful GPUs for LLM training, inference, computer vision, generative workloads, and 3D rendering —...

    Cost / License

    • Paid
    • Proprietary

    Platforms

    • Online
    search page
    Page with opensource  hosted LLMs
    GPU Burst feature
  11. MiniMax Platform is a versatile AI ecosystem offering advanced models for text, speech, video, and music generation, optimized for coding, creative expression, and immersive interaction.

    Cost / License

    • Freemium
    • Proprietary

    Platforms

    • Mac
    • Windows
    • Online
    Minimax Platform screenshot 1
    Minimax Platform screenshot 1
    Minimax Platform screenshot 2
    +1
    Minimax Platform screenshot 3
    12 alternatives
  12. GMI Cloud icon
     Like

    You’re about to supercharge your AI models with lightning-fast inference. Inference Engine makes it easy to scale, and optimize your models for real-time performance. Let’s get started—your AI is ready to shine.

    Cost / License

    • Paid
    • Proprietary

    Platforms

    • Online
    GMI Cloud screenshot 1
    GMI Cloud screenshot 1
    GMI Cloud screenshot 2
    +1
    GMI Cloud screenshot 3
    23 alternatives
  13. Lyrio.ai icon
     Like

    Lyrio.ai is your ultimate tool for structured communication. From multithreading to pinned messages, folders, and multi-LLMs, Lyrio.ai helps you manage conversations with ease. Whether you’re handling research, professional projects, or daily tasks, this AI-powered assistant is...

    Cost / License

    • Freemium
    • Proprietary

    Application type

    Platforms

    • Online
    Your Personal Organized AI Assistant
    Threads
    Multi LLMs
    8 alternatives