by nbonamy
A desktop AI assistant that bridges dozens of LLM, image, video, speech, and search providers, offering chat, generative media, RAG, shortcuts, and extensible plugins directly from the OS.
Witsy delivers a universal Model Context Protocol (MCP) client for desktop environments, letting users interact with any OpenAI‑compatible LLM as well as native services like Ollama, Anthropic, Gemini, Groq, MistralAI, DeepSeek, and many others. It centralises chat, image/video creation, speech‑to‑text, text‑to‑speech, web search, and document‑based retrieval in a single Electron‑based UI.
brew install --cask witsy
), or build from source with npm install && npm start
.Scenario | How Witsy helps |
---|---|
Developer assistance | Generate code snippets, run Python plugins, fetch docs, and paste directly into IDEs. |
Content creation | Produce images, videos, and audio narrations for blogs or social posts without leaving the desktop. |
Research & analysis | Attach PDFs, DOCX, or codebases, let the assistant summarise, extract insights, and answer questions. |
Productivity shortcuts | Transform highlighted text into commands (e.g., generate Linux commands, rewrite emails, translate on‑the‑fly). |
Accessibility | Real‑time speech‑to‑text dictation and text‑to‑speech playback for hands‑free interaction. |
Multi‑model experimentation | Switch between OpenAI, Anthropic, Ollama, Groq, etc., to compare outputs instantly. |
Q: Do I need to pay for every provider? A: Only the services you enable require a paid API key. Ollama runs locally for free.
Q: Can I run Witsy offline? A: Yes, with locally hosted models via Ollama (including embeddings) and Whisper for STT.
Q: How does Witsy store my data? A: All conversation history, settings, and document repositories are saved locally in a SQLite‑based store (future work) and can be backed up or restored via the UI.
Q: Is there a way to create my own custom command? A: Open Settings → Commands, add a new entry, define the prompt template, and assign a shortcut.
Q: What platforms are supported? A: macOS (including Homebrew cask), Windows, and Linux via the downloadable binaries.
Witsy’s Next Chapter: New Stewardship and License Update #386
Download Witsy from witsyai.com or from the releases page.
On macOS you can also brew install --cask witsy
.
Witsy is a BYOK (Bring Your Own Keys) AI application: it means you need to have API keys for the LLM providers you want to use. Alternatively, you can use Ollama to run models locally on your machine for free and use them in Witsy.
It is the first of very few (only?) universal MCP clients:Witsy allows you to run MCP servers with virtually any LLM!
Capability | Providers |
---|---|
Chat | OpenAI, Anthropic, Google (Gemini), xAI (Grok), Meta (Llama), Ollama, LM Studio, MistralAI, DeepSeek, OpenRouter, Groq, Cerebras, Azure OpenAI, any provider who supports the OpenAI API standard |
Image Creation | OpenAI (DALL-E), Google (Imagen), xAI (Grok), Replicate, fal.ai, HuggingFace, Stable Diffusion WebUI |
Video Creation | Replicate, fal.ai |
Text-to-Speech | OpenAI, ElevenLabs, Groq |
Speech-to-Text | OpenAI (Whisper), fal.ai, Fireworks.ai, Gladia, Groq, nVidia, Speechmatics, Local Whisper, Soniox (realtime and async) any provider who supports the OpenAI API standard |
Search Engines | Tavily, Brave, Exa, Local Google Search |
MCP Repositories | Smithery.ai |
Embeddings | OpenAI, Ollama |
Non-exhaustive feature list:
Generate content in any application:
On Mac, you can define an expert that will automatically be triggered depending on the foreground application. For instance, if you have an expert used to generate linux commands, you can have it selected if you trigger Prompt Anywhere from the Terminal application!
AI commands are quick helpers accessible from a shortcut that leverage LLM to boost your productivity:
You can also create custom commands with the prompt of your liking!
Commands inspired by https://the.fibery.io/@public/Public_Roadmap/Roadmap_Item/AI-Assistant-via-ChatGPT-API-170.
From https://github.com/f/awesome-chatgpt-prompts.
https://www.youtube.com/watch?v=czcSbG2H-wg
You can connect each chat with a document repository: Witsy will first search for relevant documents in your local files and provide this info to the LLM. To do so:
You can transcribe audio recorded on the microphone to text. Transcription can be done using a variety of state of the art speech to text models (which require API key) or using local Whisper model (requires download of large files).
Currently Witsy supports the following speech to text models:
Witsy supports quick shortcuts, so your transcript is always only one button press away.
Once the text is transcribed you can:
https://www.youtube.com/watch?v=vixl7I07hBk
You can download a binary from from witsyai.com, from the releases page or build yourself:
npm install
npm start
To use OpenAI, Anthropic, Google or Mistral AI models, you need to enter your API key:
To use Ollama models, you need to install Ollama and download some models.
To use text-to-speech, you need an
To use Internet search you need a Tavily API key.
Please log in to share your review and rating for this MCP.
Discover more MCP servers with similar functionality and use cases
by sooperset
MCP Atlassian is a Model Context Protocol (MCP) server that integrates AI assistants with Atlassian products like Confluence and Jira. It enables AI to automate tasks, search for information, and manage content within Atlassian ecosystems.
by ggozad
Interact with Ollama models through an intuitive terminal UI, supporting persistent chats, system prompts, model parameters, and MCP tools integration.
by GongRzhe
Provides tools for creating, editing, and enhancing PowerPoint presentations through a comprehensive set of MCP operations powered by python-pptx.
by GongRzhe
Creates, reads, and manipulates Microsoft Word documents through a standardized interface for AI assistants, enabling rich editing, formatting, and analysis capabilities.
by GongRzhe
Gmail-MCP-Server is a Model Context Protocol (MCP) server that integrates Gmail functionalities into AI assistants like Claude Desktop. It enables natural language interaction for email management, supporting features like sending, reading, and organizing emails.
by nspady
google-calendar-mcp is a Model Context Protocol (MCP) server that integrates Google Calendar with AI assistants. It enables AI assistants to manage Google Calendar events, including creating, updating, deleting, and searching for events.
by runebookai
Provides a desktop interface to chat with local or remote LLMs, schedule tasks, and integrate Model Context Protocol servers without coding.
by vivekVells
mcp-pandoc is a Model Context Protocol (MCP) server designed for seamless document format conversion using Pandoc, supporting a wide range of formats like Markdown, HTML, PDF, DOCX, and more.
by abhiz123
An MCP (Model Context Protocol) server implementation that integrates Claude with Todoist, enabling natural language task management.