

Foundry Local
Run AI models locally on your device. Foundry Local provides on-device inference with complete data privacy, no Azure subscription required.
Cost / License
- Free
- Open Source
Platforms
- Windows
- Mac
Features
- AI-Powered
Foundry Local News & Activities
Recent activities
POX added Foundry Local as alternative to QVAC Workbench- POX added Foundry Local as alternative to LocalGPT
- niksavc liked Foundry Local
- mer30hamid added Foundry Local
- mer30hamid added Foundry Local as alternative to Jan.ai, Ollama, Open WebUI and AnythingLLM
Foundry Local information
Featured in Lists
Wake up the NPU on your device
What is Foundry Local?
Foundry Local is an on-device AI inference solution that provides performance, privacy, customization, and cost benefits. It integrates with your workflows and applications through a CLI, SDK, and REST API. Key features
On-Device inference: Run models locally to reduce costs and keep data on your device.
Model customization: Select a preset model or use your own to meet specific needs.
Cost efficiency: Use existing hardware to eliminate recurring cloud costs and make AI more accessible.
Seamless integration: Integrate with your apps through the SDK, API endpoints, or CLI, and scale to Microsoft Foundry as your needs grow.
Use cases
Foundry Local is ideal when you need to:
Keep sensitive data on your device Operate in limited or offline environments Reduce cloud inference costs Get low latency AI responses for real-time applications Experiment with AI models before you deploy to the cloud

