by vectara
Provides fast, reliable retrieval‑augmented generation and semantic search via the Model Context Protocol, allowing agents to query Vectara corpora and receive generated answers together with source passages.
Vectara MCP exposes two primary tools – ask_vectara and search_vectara – that let AI agents perform RAG queries or plain semantic searches against Vectara indexes. The server implements the Model Context Protocol (MCP), so any MCP‑compatible client (e.g., Claude Desktop) can call these tools without additional glue code.
pip install vectara-mcp.uv tool run vectara-mcp).claude_desktop_config.json:{
"mcpServers": {
"Vectara": {
"command": "uv",
"args": ["tool", "run", "vectara-mcp"]
}
}
}
Q: Do I need a Vectara account? A: Yes, you must have a Vectara API key and at least one corpus key to use the tools.
Q: Can I run multiple corpora simultaneously?
A: Provide a list of corpus keys in the corpus_keys argument; the server will search across all supplied corpora.
Q: What language models are used for generation?
A: The default generation preset (vectara-summary-table-md-query-ext-jan-2025-gpt-4o) leverages a Vectara‑hosted LLM; you can specify a different preset via the generation_preset_name argument.
Q: How do I change the number of context sentences?
A: Use the optional n_sentences_before and n_sentences_after parameters when calling the tool.
Q: Is there a limit on the number of returned search results?
A: The max_used_search_results argument caps the number of results used for generation (default 10).
🔌 Compatible with Claude Desktop, and any other MCP Client!
Vectara MCP is also compatible with any MCP client
The Model Context Protocol (MCP) is an open standard that enables AI systems to interact seamlessly with various data sources and tools, facilitating secure, two-way connections.
Vectara-MCP provides any agentic application with access to fast, reliable RAG with reduced hallucination, powered by Vectara's Trusted RAG platform, through the MCP protocol.
You can install the package directly from PyPI:
pip install vectara-mcp
# Start server with secure HTTP transport (DEFAULT)
python -m vectara_mcp
# Server running at http://127.0.0.1:8000 with authentication enabled
# For Claude Desktop or local development (less secure)
python -m vectara_mcp --stdio
# ⚠️ Warning: STDIO transport is less secure. Use only for local development.
# Custom host and port
python -m vectara_mcp --host 0.0.0.0 --port 8080
# SSE transport mode
python -m vectara_mcp --transport sse --path /sse
# Disable authentication (DANGEROUS - dev only)
python -m vectara_mcp --no-auth
--stdio flag# Required
export VECTARA_API_KEY="your-api-key"
# Optional
export VECTARA_AUTHORIZED_TOKENS="token1,token2" # Additional auth tokens
export VECTARA_ALLOWED_ORIGINS="http://localhost:*,https://app.example.com"
export VECTARA_TRANSPORT="http" # Default transport mode
export VECTARA_AUTH_REQUIRED="true" # Enforce authentication
When using HTTP or SSE transport, authentication is required by default:
# Using curl with bearer token
curl -H "Authorization: Bearer $VECTARA_API_KEY" \
-H "Content-Type: application/json" \
-X POST http://localhost:8000/call/ask_vectara \
-d '{"query": "What is Vectara?", "corpus_keys": ["my-corpus"]}'
# Using X-API-Key header (alternative)
curl -H "X-API-Key: $VECTARA_API_KEY" \
http://localhost:8000/sse
# ⚠️ NEVER use in production
python -m vectara_mcp --no-auth
setup_vectara_api_key: Configure and validate your Vectara API key for the session (one-time setup).
Args:
Returns:
clear_vectara_api_key: Clear the stored API key from server memory.
Returns:
ask_vectara: Run a RAG query using Vectara, returning search results with a generated response.
Args:
Returns:
search_vectara: Run a semantic search query using Vectara, without generation.
Args:
Returns:
correct_hallucinations: Identify and correct hallucinations in generated text using Vectara's VHC (Vectara Hallucination Correction) API.
Args:
Returns:
eval_factual_consistency: Evaluate the factual consistency of generated text against source documents using Vectara's dedicated factual consistency evaluation API.
Args:
Returns:
Note: API key must be configured first using setup_vectara_api_key tool or VECTARA_API_KEY environment variable.
To use with Claude Desktop, update your configuration to use STDIO transport:
{
"mcpServers": {
"Vectara": {
"command": "python",
"args": ["-m", "vectara_mcp", "--stdio"],
"env": {
"VECTARA_API_KEY": "your-api-key"
}
}
}
}
Or using uv:
{
"mcpServers": {
"Vectara": {
"command": "uv",
"args": ["tool", "run", "vectara-mcp", "--stdio"]
}
}
}
Note: Claude Desktop requires STDIO transport. While less secure than HTTP, it's acceptable for local desktop use.
Once the installation is complete, and the Claude desktop app is configured, you must completely close and re-open the Claude desktop app to see the Vectara-mcp server. You should see a hammer icon in the bottom left of the app, indicating available MCP tools, you can click on the hammer icon to see more detail on the Vectara-search and Vectara-extract tools.
Now claude will have complete access to the Vectara-mcp server, including all six Vectara tools.
First-time setup (one-time per session):
setup-vectara-api-key
API key: [your-vectara-api-key]
After setup, use any tools without exposing your API key:
ask-vectara
Query: Who is Amr Awadallah?
Corpus keys: ["your-corpus-key"]
search-vectara
Query: events in NYC?
Corpus keys: ["your-corpus-key"]
correct-hallucinations
Generated text: [text to check]
Documents: ["source1", "source2"]
eval-factual-consistency
Generated text: [text to evaluate]
Documents: ["reference1", "reference2"]
--no-auth for local testingVECTARA_ALLOWED_ORIGINS to restrict accessVECTARA_API_KEY and VECTARA_AUTHORIZED_TOKENSSee SECURITY.md for detailed security guidelines.
For issues, questions, or contributions, please visit: https://github.com/vectara/vectara-mcp
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by exa-labs
Provides real-time web search capabilities to AI assistants via a Model Context Protocol server, enabling safe and controlled access to the Exa AI Search API.
by perplexityai
Enables Claude and other MCP‑compatible applications to perform real‑time web searches through the Perplexity (Sonar) API without leaving the MCP ecosystem.
by MicrosoftDocs
Provides semantic search and fetch capabilities for Microsoft official documentation, returning content in markdown format via a lightweight streamable HTTP transport for AI agents and development tools.
by elastic
Enables natural‑language interaction with Elasticsearch indices via the Model Context Protocol, exposing tools for listing indices, fetching mappings, performing searches, running ES|QL queries, and retrieving shard information.
by graphlit
Enables integration between MCP clients and the Graphlit platform, providing ingestion, extraction, retrieval, and RAG capabilities across a wide range of data sources and connectors.
by ihor-sokoliuk
Provides web search capabilities via the SearXNG API, exposing them through an MCP server for seamless integration with AI agents and tools.
by mamertofabian
Fast cross‑platform file searching leveraging the Everything SDK on Windows, Spotlight on macOS, and locate/plocate on Linux.
by spences10
Provides unified access to multiple search engines, AI response tools, and content processing services through a single Model Context Protocol server.
by cr7258
Provides Elasticsearch and OpenSearch interaction via Model Context Protocol, enabling document search, index management, cluster monitoring, and alias operations.
{
"mcpServers": {
"Vectara": {
"command": "uv",
"args": [
"tool",
"run",
"vectara-mcp"
]
}
}
}claude mcp add Vectara uv tool run vectara-mcp