Witsy

Witsy

By nbonamy GitHub

Witsy: desktop AI assistant

vuejs mcp
Overview

What is Witsy?

Witsy is a desktop AI assistant that allows users to leverage various AI models for productivity and creativity. It supports multiple AI providers and enables users to run models locally or through API keys.

How to use Witsy?

To use Witsy, download it from witsyai.com or the releases page. Set up your API keys for the desired AI models, or use local models via Ollama.

Key features of Witsy?

  • Supports multiple AI models including OpenAI, Ollama, and Anthropic.
  • Chat completion with vision model support.
  • Text-to-image and text-to-video generation.
  • Scratchpad for interactive content creation.
  • AI commands for productivity enhancements.
  • Long-term memory plugin for improved relevance.
  • Transcription and dictation capabilities.
  • Local file interaction for document-based queries.

Use cases of Witsy?

  1. Generating creative content across applications.
  2. Assisting with coding tasks and generating code snippets.
  3. Transcribing audio to text for documentation.
  4. Enhancing productivity with AI commands in various applications.

FAQ from Witsy?

  • Can Witsy work without API keys?

Yes, you can use local models via Ollama without API keys.

  • Is Witsy free to use?

Witsy is free to download and use, but some features may require API keys from third-party providers.

  • What types of AI models does Witsy support?

Witsy supports a variety of models including those from OpenAI, Anthropic, and local models via Ollama.

Content

Witsy Logo

Witsy
Desktop AI Assistant

Version Badge Downloads Badge Test Badge Coverage Badge

Downloads

Download Witsy from witsyai.com or from the releases page.

What is Witsy?

Witsy is a BYOK (Bring Your Own Keys) AI application: it means you need to have API keys for the LLM providers you want to use. Alternatively, you can use Ollama to run models locally on your machine for free and use them in Witsy.

Non-exhaustive feature list:

  • OpenAI, Ollama, Anthropic, MistralAI, Google, xAI, Azure, OpenRouter, DeepSeek, Groq and Cerebras models supported
  • Connect other providers (together, siliconflow, fireworks...) through OpenAI compatibility layer
  • Chat completion with vision models support (describe an image)
  • Text-to-image and text-to video with OpenAI, Google, xAI, Replicate, fal.ai and HuggingFace
  • Image-to-image (image editing) and image-to-video with Google, Replicate and fal.ai
  • Scratchpad to interactively create the best content with any model!
  • Prompt anywhere allows to generate content directly in any application
  • AI commands runnable on highlighted text in almost any application
  • Experts prompts to specialize your bot on a specific topic
  • LLM plugins to augment LLM: execute python code, search the Internet...
  • Anthropic MCP server support
  • Long-term memory plugin to increase relevance of LLM answers
  • Read aloud of assistant messages (requires OpenAI or ElevenLabs API key)
  • Read aloud of any text in other applications (requires OpenAI or ElevenLabs API key)
  • Chat with your local files and documents (RAG)
  • Transcription/Dictation (Speech-to-Text)
  • Realtime Chat aka Voice Mode
  • Anthropic Computer Use support
  • Local history of conversations (with automatic titles)
  • Formatting and copy to clipboard of generated code
  • Conversation PDF export
  • Image copy and download

     

Prompt Anywhere

Generate content in any application:

  • From any editable content in any application
  • Hit the Prompt anywhere shortcut (Shift+Control+Space / ^⇧Space)
  • Enter your prompt in the window that pops up
  • Watch Witsy enter the text directly in your application!

On Mac, you can define an expert that will automatically be triggered depending on the foreground application. For instance, if you have an expert used to generate linux commands, you can have it selected if you trigger Prompt Anywhere from the Terminal application!

AI Commands

AI commands are quick helpers accessible from a shortcut that leverage LLM to boost your productivity:

  • Select any text in any application
  • Hit the AI command shorcut (Alt+Control+Space / ⌃⌥Space)
  • Select one of the commands and let LLM do their magic!

You can also create custom commands with the prompt of your liking!

     

Commands inspired by https://the.fibery.io/@public/Public_Roadmap/Roadmap_Item/AI-Assistant-via-ChatGPT-API-170.

Experts

From https://github.com/f/awesome-chatgpt-prompts.

Scratchpad

https://www.youtube.com/watch?v=czcSbG2H-wg

Chat with your documents (RAG)

You can connect each chat with a document repository: Witsy will first search for relevant documents in your local files and provide this info to the LLM. To do so:

  • Click on the database icon on the left of the prompt
  • Click Manage and then create a document repository
  • OpenAI Embedding require on API key, Ollama requires an embedding model
  • Add documents by clicking the + button on the right hand side of the window
  • Once your document repository is created, click on the database icon once more and select the document repository you want to use. The icon should turn blue

Transcription / Dictation (Speech-to-Text)

You can transcribe audio recorded on the microphone to text. Transcription can be done using OpenAI Whisper online model (requires API key) or using local Whisper model (requires download of large files). Once the text is transcribed you can:

  • Copy it to your clipboard
  • Insert it in the application that was running before you activated the dictation

Anthropic Computer Use

https://www.youtube.com/watch?v=vixl7I07hBk

Setup

You can download a binary from from witsyai.com, from the releases page or build yourself:

npm install
npm start

Prerequisites

To use OpenAI, Anthropic, Google or Mistral AI models, you need to enter your API key:

To use Ollama models, you need to install Ollama and download some models.

To use text-to-speech, you need an OpenAI API key.

To use Internet search you need a Tavily API key.

  

TODO

  • Proper database (SQLite3) storage (??)

WIP

DONE

  • Azure AI support
  • Brave Search plugin
  • Allow user-input models for embeddings
  • User defined parameters for custom engines
  • Direct speech-to-text checbox
  • Quick access buttons on home
  • fal.ai support (speech-to-text, text-to-image and text-to-video)
  • Debug console
  • Design Studio
  • i18n
  • Mermaid diagram rendering
  • Smithery.ai MCP integration
  • Model Context Protocol
  • Local Web Search
  • Model defaults
  • Speech-to-text language
  • Model parameters (temperature...)
  • Favorite models
  • ElevenLabs Text-to-Speech
  • Custom engines (OpenAI compatible)
  • Long-term memory plugin
  • OpenRouter support
  • DeepSeek support
  • Folder mode
  • All instructions customization
  • Fork chat (with optional LLM switch)
  • Realtime chat
  • Replicate video generation
  • Together.ai compatibility
  • Gemini 2.0 Flash support
  • Groq LLama 3.3 support
  • xAI Grok Vision Model support
  • Ollama function-calling
  • Replicate image generation
  • AI Commands redesign
  • Token usage report
  • OpenAI o1 models support
  • Groq vision support
  • Image resize option
  • Llama 3.2 vision support
  • YouTube plugin
  • RAG in Scratchpad
  • Hugging face image generation
  • Show prompt used for image generation
  • Redesigned Prompt window
  • Anthropic Computer Use
  • Auto-update refactor (still not Windows)
  • Dark mode
  • Conversation mode
  • Google function calling
  • Anthropic function calling
  • Scratchpad
  • Dictation: OpenAI Whisper + Whisper WebGPU
  • Auto-select expert based on foremost app (Mac only)
  • Cerebras support
  • Local files RAG
  • Groq model update (8-Sep-2024)
  • PDF Export of chats
  • Prompts renamed to Experts. Now editable.
  • Read aloud
  • Import/Export commands
  • Anthropic Sonnet 3.5
  • Ollama base URL as settings
  • OpenAI base URL as settings
  • DALL-E as tool
  • Google Gemini API
  • Prompt anywhere
  • Cancel commands
  • GPT-4o support
  • Different default engine/model for commands
  • Text attachments (TXT, PDF, DOCX, PPTX, XLSX)
  • MistralAI function calling
  • Auto-update
  • History date sections
  • Multiple selection delete
  • Search
  • Groq API
  • Custom prompts
  • Sandbox & contextIsolation
  • Application Menu
  • Prompt history navigation
  • Ollama model pull
  • macOS notarization
  • Fix when long text is highlighted
  • Shortcuts for AI commands
  • Shift to switch AI command behavior
  • User feedback when running a tool
  • Download internet content plugin
  • Tavily Internet search plugin
  • Python code execution plugin
  • LLM Tools supprt (OpenAI only)
  • Mistral AI API integration
  • Latex rendering
  • Anthropic API integration
  • Image generation as b64_json
  • Text-to-speech
  • Log file (electron-log)
  • Conversation language settings
  • Paste image in prompt
  • Run commands with default models
  • Models refresh
  • Edit commands
  • Customized commands
  • Conversation menu (info, save...)
  • Conversation depth setting
  • Save attachment on disk
  • Keep running in system tray
  • Nicer icon (still temporary)
  • Rename conversation
  • Copy/edit messages
  • New chat window for AI command
  • AI Commands with shortcut
  • Auto-switch to vision model
  • Run at login
  • Shortcut editor
  • Chat font size settings
  • Image attachment for vision
  • Stop response streaming
  • Save/Restore window position
  • Ollama support
  • View image full screen
  • Status/Tray bar icon + global shortcut to invoke
  • Chat themes
  • Default instructions in settings
  • Save DALL-E images locally (and delete properly)
  • OpenAI links in settings
  • Copy code button
  • Chat list ordering
  • OpenAI model choice
  • CSS variables
No tools information available.
School MCP
School MCP by 54yyyu

A Model Context Protocol (MCP) server for academic tools, integrating with Canvas and Gradescope platforms.

canvas mcp
View Details
repo-template
repo-template by loonghao

A Model Context Protocol (MCP) server for Python package intelligence, providing structured queries for PyPI packages and GitHub repositories. Features include dependency analysis, version tracking, and package metadata retrieval for LLM interactions.

-

google-calendar mcp
View Details
strava-mcp
strava-mcp by jeremysilva1098

MCP server for strava

strava mcp
View Details

Model Context Protocol (MCP) server implementation for Rhinoceros/Grasshopper integration, enabling AI models to interact with parametric design tools

grasshopper mcp
View Details

MCP configuration to connect AI agent to a Linux machine.

security mcp
View Details

AI assistant built with Streamlit, NVIDIA NIM (LLaMa 3.3:70B) / Ollama, and Model Control Protocol (MCP).

python mcp
View Details