

LLM Hub
LLM Hub is an open-source Android app for on-device LLM chat and image generation. It's optimized for mobile usage (CPU/GPU/NPU acceleration) and supports multiple model formats so you can run powerful models locally and privately.
Cost / License
- Free
- Open Source (MIT)
Platforms
- Android
- Android Tablet
Features
Properties
- Privacy focused
Features
- Hardware Accelerated
- OCR
- No Tracking
- Ad-free
- No registration required
- Works Offline
- Dark Mode
- Text to Speech
- Text to Image Generation
- GPU Acceleration
- speech transcription
- AI-Powered
- Speech Recognition
- Speech to text
- Translator
- Offline
- Material design
Tags
- on-device-ai
- asr
- phi4
- gemma3
- npu
- qnn
- Stable Diffusion
- transcription
- gguf
- stt
- RAG
- mediapipe
- litert
- AI
- retrieval-augmented-generation
- llama
- mnn
- multimodal
- llm-inference
- local-ai
- onnx
- scamdetector
- gemma3n
LLM Hub News & Activities
Recent activities
- bugmenot added LLM Hub
bugmenot added LLM Hub as alternative to RWKV Chat, NexaSDK, Lemonade Server and AI Playground
LLM Hub information
What is LLM Hub?
LLM Hub is an open-source Android app for on-device LLM chat and image generation. It's optimized for mobile usage (CPU/GPU/NPU acceleration) and supports multiple model formats so you can run powerful models locally and privately.
? SIX AI TOOLS
📝 CHAT Multi-turn conversations with RAG memory, web search, TTS auto-readout, and multimodal input (text, images, audio)
?? WRITING AID Summarize, expand, rewrite, improve grammar, or generate code from descriptions
🎨 IMAGE GENERATOR Create images from text prompts using Stable Diffusion 1.5 with swipeable gallery for variations
🌍 TRANSLATOR Translate text, images (OCR), and audio across 50+ languages - works offline
🎙? TRANSCRIBER Convert speech to text with on-device processing
🛡? SCAM DETECTOR Analyze messages and images for phishing with risk assessment.
🔐 PRIVACY & SECURITY
• 100% on-device processing - no internet required for inference • Zero data collection - conversations never leave your device • No accounts, no tracking - completely private • Open-source - fully transparent
? ADVANCED FEATURES
• GPU/NPU acceleration for fast performance • Text-to-Speech with auto-readout • RAG with global memory for enhanced responses • Import custom models (.task, .litertlm, .mnn, .gguf) • Direct downloads from HuggingFace • 16 language interfaces
Quick Start
- Download from Google Play or build from source
- Open Settings ? Download Models ? Download or Import a model
- Select a model and start chatting or generating images
Supported Model Families (summary)
• Gemma (LiteRT Task) • Llama (Task + GGUF variants) • Phi (LiteRT LM) • LiquidAI LFM (LFM 2.5 1.2B + LFM VL 1.6B vision-enabled) • Ministral / Mistral family (GGUF / ONNX) • IBM Granite (GGUF)
Model Formats
• Task / LiteRT (.task): MediaPipe/LiteRT optimized models (GPU/NPU capable) • LiteRT LM (.litertlm): LiteRT language models • GGUF (.gguf): Quantized models — CPU inference powered by Nexa SDK; some vision-capable GGUF models require an additional mmproj vision project file • ONNX (.onnx): Cross-platform model runtime
Importing models
• Settings ? Download Models ? Import Model ? choose .task, .litertlm, .mnn, .gguf, or .onnx • The full model list and download links live in app/src/.../data/ModelData.kt (do not exhaustively list variants in the README)
Technology
• Kotlin + Jetpack Compose (Material 3) • LLM Runtime: MediaPipe, LiteRT, Nexa SDK • Image Gen: MNN / Qualcomm QNN • Quantization: INT4/INT8
Acknowledgments
• Nexa SDK — GGUF model inference support (credit shown in-app About) ? • Google, Meta, Microsoft, IBM, LiquidAI, Mistral, HuggingFace — model and tooling contributions







