by nbonamy
A desktop AI assistant that integrates with dozens of LLM providers, offering chat, multimodal generation, RAG, shortcuts, and extensible commands across the operating system.
Witsy provides a universal client for Model Context Protocol (MCP) servers, enabling users to interact with any LLM through a single desktop application. It supports a broad range of AI capabilities—chat, vision, image/video generation, text‑to‑speech, speech‑to‑text, internet search, embeddings, and document‑based retrieval—while keeping all API keys under the user's control (BYOK).
brew install --cask witsy
on macOS), or build from source:
npm install
npm start
Q: Do I need to pay for the models? A: Only for the providers you enable. Ollama can run locally for free; all other services require their own API keys.
Q: Can I use Witsy offline? A: Yes, with locally hosted models via Ollama or any OpenAI‑compatible server you run yourself.
Q: How are my API keys stored? A: Keys are saved locally in encrypted form and never transmitted to any third‑party server.
Q: Is there a macOS notarization? A: Yes, the macOS binary is notarized and can be installed via Homebrew.
Q: How do I add a new AI Command? A: Open Settings → AI Commands → "Add", define the prompt template and assign a shortcut.
Q: What is the difference between Experts and Prompts? A: Experts are editable prompt collections that can be auto‑selected based on the foreground app; they replace the older "Prompt" terminology.
Witsy’s Next Chapter: New Stewardship and License Update #386
Download Witsy from witsyai.com or from the releases page.
On macOS you can also brew install --cask witsy
.
Witsy is a BYOK (Bring Your Own Keys) AI application: it means you need to have API keys for the LLM providers you want to use. Alternatively, you can use Ollama to run models locally on your machine for free and use them in Witsy.
It is the first of very few (only?) universal MCP clients:Witsy allows you to run MCP servers with virtually any LLM!
Capability | Providers |
---|---|
Chat | OpenAI, Anthropic, Google (Gemini), xAI (Grok), Meta (Llama), Ollama, LM Studio, MistralAI, DeepSeek, OpenRouter, Groq, Cerebras, Azure OpenAI, any provider who supports the OpenAI API standard |
Image Creation | OpenAI (DALL-E), Google (Imagen), xAI (Grok), Replicate, fal.ai, HuggingFace, Stable Diffusion WebUI |
Video Creation | Replicate, fal.ai |
Text-to-Speech | OpenAI, ElevenLabs, Groq |
Speech-to-Text | OpenAI (Whisper), fal.ai, Fireworks.ai, Gladia, Groq, nVidia, Speechmatics, Local Whisper, Soniox (realtime and async) any provider who supports the OpenAI API standard |
Search Engines | Tavily, Brave, Exa, Local Google Search |
MCP Repositories | Smithery.ai |
Embeddings | OpenAI, Ollama |
Non-exhaustive feature list:
Generate content in any application:
On Mac, you can define an expert that will automatically be triggered depending on the foreground application. For instance, if you have an expert used to generate linux commands, you can have it selected if you trigger Prompt Anywhere from the Terminal application!
AI commands are quick helpers accessible from a shortcut that leverage LLM to boost your productivity:
You can also create custom commands with the prompt of your liking!
Commands inspired by https://the.fibery.io/@public/Public_Roadmap/Roadmap_Item/AI-Assistant-via-ChatGPT-API-170.
From https://github.com/f/awesome-chatgpt-prompts.
https://www.youtube.com/watch?v=czcSbG2H-wg
You can connect each chat with a document repository: Witsy will first search for relevant documents in your local files and provide this info to the LLM. To do so:
You can transcribe audio recorded on the microphone to text. Transcription can be done using a variety of state of the art speech to text models (which require API key) or using local Whisper model (requires download of large files).
Currently Witsy supports the following speech to text models:
Witsy supports quick shortcuts, so your transcript is always only one button press away.
Once the text is transcribed you can:
https://www.youtube.com/watch?v=vixl7I07hBk
You can download a binary from from witsyai.com, from the releases page or build yourself:
npm install
npm start
To use OpenAI, Anthropic, Google or Mistral AI models, you need to enter your API key:
To use Ollama models, you need to install Ollama and download some models.
To use text-to-speech, you need an
To use Internet search you need a Tavily API key.
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by GLips
Provides Figma layout and styling information to AI coding agents, enabling one‑shot implementation of designs in any framework.
by idosal
Provides a remote Model Context Protocol server that transforms any public GitHub repository into an up‑to‑date documentation hub, enabling AI assistants to fetch live code and docs, dramatically reducing hallucinations.