Osaurus icon
Osaurus icon

Osaurus

Native, Apple Silicon–only local LLM server. Similar to Ollama, but built on Apple's MLX for maximum performance on M-series chips. SwiftUI app + SwiftNIO server with OpenAI-compatible endpoints.

Osaurus screenshot 1

Cost / License

  • Free
  • Open Source

Platforms

  • Mac
-
No reviews
1like
0comments
0news articles

Features

Suggest and vote on features
  1.  Dark Mode
  2.  Sits in the MenuBar
  3.  AI-Powered
  4.  OpenAI integration
  5.  Apple Silicon support

 Tags

  • local-ai
  • ai-models
  • ai-model-integration
  • openai-api
  • llm-integration
  • large-language-model-tool

Osaurus News & Activities

Highlights All activities

Recent activities

Show all activities

Osaurus information

  • Developed by

    US flagDinoki, LLC
  • Licensing

    Open Source (MIT) and Free product.
  • Written in

  • Alternatives

    41 alternatives listed
  • Supported Languages

    • English

AlternativeTo Category

AI Tools & Services

GitHub repository

  •  2,627 Stars
  •  106 Forks
  •  15 Open Issues
  •   Updated  
View on GitHub

Popular alternatives

View all
Osaurus was added to AlternativeTo by Paul on and this page was last updated .
No comments or reviews, maybe you want to be first?
Post comment/review

What is Osaurus?

Native, Apple Silicon–only local LLM server. Similar to Ollama icon Ollama, but built on Apple's MLX for maximum performance on M-series chips. SwiftUI app + SwiftNIO server with OpenAI-compatible endpoints.

Created by Dinoki icon Dinoki, a fully native desktop AI assistant and companion.

Highlights:

  • Native MLX runtime: Optimized for Apple Silicon using MLX/MLXLLM
  • Apple Silicon only: Designed and tested for M-series Macs
  • OpenAI API compatible: /v1/models and /v1/chat/completions (stream and non-stream)
  • Function/Tool calling: OpenAI-style tools + tool_choice, with tool_calls parsing and streaming deltas
  • Chat templates: Uses model-provided Jinja chat_template with BOS/EOS, with smart fallback
  • Session reuse (KV cache): Faster multi-turn chats via session_id
  • Fast token streaming: Server-Sent Events for low-latency output
  • Model manager UI: Browse, download, and manage MLX models from mlx-community
  • System resource monitor: Real-time CPU and RAM usage visualization
  • Self-contained: SwiftUI app with an embedded SwiftNIO HTTP server

Official Links