Vectara MCP Server

Vectara MCP Server

By vectara GitHub

Vectara-MCP provides any Agentic AI application with access to fast, reliable RAG with reduced hallucinations, powered by Vectara's Trusted RAG platform, through the MCP protocol.

vectara RAG
Overview

What is Vectara MCP?

Vectara MCP is a server that provides agentic AI applications with access to fast and reliable Retrieval-Augmented Generation (RAG) capabilities, powered by Vectara's Trusted RAG platform through the Model Context Protocol (MCP).

How to use Vectara MCP?

To use Vectara MCP, configure it with your Claude Desktop app by adding the necessary server details to your configuration file. After setup, you can utilize the ask_vectara and search_vectara tools to run queries and retrieve information.

Key features of Vectara MCP?

  • Fast and reliable RAG capabilities with reduced hallucinations.
  • Compatibility with any MCP client, including Claude Desktop.
  • Tools for running RAG queries and semantic searches.

Use cases of Vectara MCP?

  1. Enhancing AI applications with reliable data retrieval.
  2. Running complex queries to extract information from various data sources.
  3. Integrating with other AI systems for improved performance.

FAQ from Vectara MCP?

  • What is the Model Context Protocol (MCP)?

MCP is an open standard that allows AI systems to interact with various data sources and tools securely.

  • Is Vectara MCP free to use?

Yes, Vectara MCP is available for free to users.

  • How do I configure Vectara MCP with Claude Desktop?

You need to add the server details to your claude_desktop_config.json file and restart the app to see the changes.

Content

Vectara MCP Server 🚀

GitHub Repo stars

🔌 Compatible with Claude Desktop, and any other MCP Client!

Vectara MCP is also compatible with any MCP client

The Model Context Protocol (MCP) is an open standard that enables AI systems to interact seamlessly with various data sources and tools, facilitating secure, two-way connections.

Vectara-MCP provides any agentic application with access to fast, reliable RAG with reduced hallucination, powered by Vectara's Trusted RAG platform, through the MCP protocol.

Available Tools 🔧

  • ask_vectara: Run a RAG query using Vectara, returning search results with a generated response.

    Args:

    • query: str, The user query to run - required.
    • corpus_keys: list[str], List of Vectara corpus keys to use for the search - required. Please ask the user to provide one or more corpus keys.
    • api_key: str, The Vectara API key - required.
    • n_sentences_before: int, Number of sentences before the answer to include in the context - optional, default is 2.
    • n_sentences_after: int, Number of sentences after the answer to include in the context - optional, default is 2.
    • lexical_interpolation: float, The amount of lexical interpolation to use - optional, default is 0.005.
    • max_used_search_results: int, The maximum number of search results to use - optional, default is 10.
    • generation_preset_name: str, The name of the generation preset to use - optional, default is "vectara-summary-table-md-query-ext-jan-2025-gpt-4o".
    • response_language: str, The language of the response - optional, default is "eng".

    Returns:

    • The response from Vectara, including the generated answer and the search results.
  • search_vectara: Run a semantic search query using Vectara, without generation.

    Args:

    • query: str, The user query to run - required.
    • corpus_keys: list[str], List of Vectara corpus keys to use for the search - required. Please ask the user to provide one or more corpus keys.
    • api_key: str, The Vectara API key - required.
    • n_sentences_before: int, Number of sentences before the answer to include in the context - optional, default is 2.
    • n_sentences_after: int, Number of sentences after the answer to include in the context - optional, default is 2.
    • lexical_interpolation: float, The amount of lexical interpolation to use - optional, default is 0.005.

    Returns:

    • The response from Vectara, including the matching search results.

Configuration with Claude Desktop ⚙️

Add to your claude_desktop_config.json:

{
  "mcpServers": {
    "Vectara": {
      "command": "npx",
      "args": ["/path/to/vectara-mcp/build/index.js"],
    }
  }
}

Replace /path/to/vectara-mcp with the actual path where you cloned the repository on your system.

Usage in Claude Desktop App 🎯

Once the installation is complete, and the Claude desktop app is configured, you must completely close and re-open the Claude desktop app to see the Vectara-mcp server. You should see a hammer icon in the bottom left of the app, indicating available MCP tools, you can click on the hammer icon to see more detial on the Vectara-search and Vectara-extract tools.

Now claude will have complete access to the Vectara-mcp server, including the ask-vectara and search-vectara tools. When you issue the tools for the first time, Claude will ask you for your Vectara api key and corpus key (or keys if you want to use multiple corpora). After you set those, you will be ready to go. Here are some examples you can try (with the Vectara corpus that includes information from our website:

Vectara RAG Examples

  1. Querying Vectara corpus:
ask-vectara Who is Amr Awadallah?
  1. Searching Vectara corpus:
search-vectara events in NYC?

Acknowledgments ✨

No tools information available.

A powerful knowledge management system that forges wisdom from experiences, insights, and best practices. Built with Qdrant vector database for efficient knowledge storage and retrieval.

mcp-local-rag
mcp-local-rag by nkapila6

"primitive" RAG-like web search model context protocol (MCP) server that runs locally. ✨ no APIs ✨

A custom MCP server with RAG capabilities and multiple search providers (Gemini 2.0 and Linkup)

Tiny Chat
Tiny Chat by to-aoki

This is an LLM application with chat functionality, featuring chat using RAG, a database, and MCP server capabilities. The UI is designed for Japanese users.

An MCP server implementation that provides tools for retrieving and processing documentation through vector search, enabling AI assistants to augment their responses with relevant documentation context.

A Model Context Protocol (MCP) server implementation for RAG

Vectara-MCP provides any Agentic AI application with access to fast, reliable RAG with reduced hallucinations, powered by Vectara's Trusted RAG platform, through the MCP protocol.

vectara RAG
View Details