TL;DR
Osaurus, an open-source Mac app, now allows users to run and switch between local and cloud AI models securely. The tool aims to enhance privacy and flexibility for individual users and businesses.
Osaurus has launched a new Mac application that enables users to run and switch between local and cloud AI models securely on their hardware, marking a significant step in local AI deployment.
Osaurus is an open-source, Apple-only AI server that allows users to connect with locally hosted AI models or cloud providers such as OpenAI and Anthropic. The application offers a user-friendly interface designed for consumers, contrasting with developer-focused tools like OpenClaw, and emphasizes security through hardware-isolated virtual sandboxes.
The platform supports a variety of models, including MiniMax M2.5, Gemma 4, Qwen3.6, GPT-OSS, Llama, and DeepSeek V4, as well as Apple’s on-device models and Liquid AI’s LFM family. It also integrates with multiple cloud services, providing a comprehensive control layer for AI workflows. Recent updates added voice capabilities, and the project has garnered over 112,000 downloads since its release nearly a year ago.
Why It Matters
This development matters because it advances the accessibility and security of local AI deployment, reducing dependence on cloud data centers and addressing privacy concerns. It also offers a flexible platform for individuals and businesses to leverage AI models without extensive technical knowledge, potentially transforming how AI is integrated into daily workflows and enterprise applications.
Mac AI model hosting software
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
Background
The rise of AI model commoditization has prompted startups to develop software layers that facilitate switching between local and cloud models. Osaurus evolved from the idea of a desktop AI companion, Dinoki, which aimed to run AI locally on Macs. The current version supports a broad range of models and cloud services, reflecting ongoing innovation in local AI hardware and software. Hardware requirements remain high—at least 64 GB of RAM for smaller models and 128 GB for larger ones—highlighting the resource-intensive nature of local AI deployment.
“You can do pretty much everything on your Mac locally, like browsing your files, accessing your browser, and system configurations. This makes Osaurus a personal AI for individuals.”
— Terence Pae, Osaurus co-founder
“The intelligence per wattage has been going up significantly, and local AI is just getting better and better—today it can run tools, write code, and even order from Amazon.”
— Terence Pae, Osaurus co-founder

A Cowboy's Guide to Setting Up Your Own Garage AI Agent: Run Local AI Models on a Budget Using Apple Silicon, Ollama, MCP, RAG & More (THE COWBOY'S GUIDE SERIES)
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
What Remains Unclear
It remains unclear how widely adopted Osaurus will become among consumers and enterprises, and how hardware requirements will evolve to make local AI more accessible. The impact on cloud AI demand and data center reliance is still speculative at this stage.
AI cloud and local model switcher Mac
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
What’s Next
Next steps include expanding user adoption, further refining features like voice capabilities, and exploring enterprise applications such as legal and healthcare sectors. The team is also participating in startup accelerators to scale deployment and partnerships.

AI VoiceWriter – Smart Dictation & AI Writing Assistant for Windows & Mac | USB Dongle & Mobile App for Voice Input, Proofreading, Rewriting & Multilingual Support
🎙️ Hands-Free Voice Typing for Windows & Mac – Powered by iOS & Android dictation technology, AI VoiceWriter…
As an affiliate, we earn on qualifying purchases.
As an affiliate, we earn on qualifying purchases.
Key Questions
Can I run Osaurus on my current Mac?
Osaurus requires Macs with at least 64 GB of RAM for smaller models and 128 GB for larger ones. Hardware specifications may limit some users currently.
Does Osaurus support all AI models and cloud providers?
It supports a wide range of models, including MiniMax, Gemma, Qwen, GPT-OSS, Llama, and DeepSeek V4, as well as Apple’s and Liquid AI’s models. It also connects to major cloud providers like OpenAI and Anthropic.
Is Osaurus suitable for enterprise use?
While primarily designed for individual users, Osaurus is exploring enterprise applications, especially in privacy-sensitive fields like healthcare and legal services.