by sirmews
Read and write records in a Pinecone vector index via Model Context Protocol, enabling semantic search and document management for Claude Desktop.
Mcp Pinecone provides Model Context Protocol endpoints that let a client (e.g., Claude Desktop) interact with a Pinecone vector index. It supports reading, writing, searching, and utility operations such as stats and document processing.
uvx install mcp-pinecone # or uv pip install mcp-pinecone
claude_desktop_config.json
:
{
"mcpServers": {
"mcp-pinecone": {
"command": "uvx",
"args": [
"--index-name",
"{your-index-name}",
"--api-key",
"{your-secret-api-key}",
"mcp-pinecone"
]
}
}
}
semantic-search
, read-document
, list-documents
, etc.) from the Claude UI to interact with your Pinecone index.pinecone-stats
tool.Q: Do I need a Pinecone account? A: Yes. Sign up at pinecone.io, create an index, and obtain an API key.
Q: Which embedding model is used? A: Embeddings are generated via Pinecone’s built‑in inference API.
Q: Can I run the server locally without Claude? A: The server runs over stdio and expects MCP‑compatible clients; you can test it with the MCP Inspector.
Q: How are documents chunked? A: A token‑based chunker similar to LangChain splits documents before embedding.
Q: How do I publish updates to PyPI?
A: Use uv build
to create distributions and uv publish
with your PyPI token.
Read and write to a Pinecone index.
The server implements the ability to read and write to a Pinecone index.
semantic-search
: Search for records in the Pinecone index.read-document
: Read a document from the Pinecone index.list-documents
: List all documents in the Pinecone index.pinecone-stats
: Get stats about the Pinecone index, including the number of records, dimensions, and namespaces.process-document
: Process a document into chunks and upsert them into the Pinecone index. This performs the overall steps of chunking, embedding, and upserting.Note: embeddings are generated via Pinecone's inference API and chunking is done with a token-based chunker. Written by copying a lot from langchain and debugging with Claude.
To install Pinecone MCP Server for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install mcp-pinecone --client claude
Recommend using uv to install the server locally for Claude.
uvx install mcp-pinecone
OR
uv pip install mcp-pinecone
Add your config as described below.
On MacOS: ~/Library/Application\ Support/Claude/claude_desktop_config.json
On Windows: %APPDATA%/Claude/claude_desktop_config.json
Note: You might need to use the direct path to uv
. Use which uv
to find the path.
Development/Unpublished Servers Configuration
"mcpServers": {
"mcp-pinecone": {
"command": "uv",
"args": [
"--directory",
"{project_dir}",
"run",
"mcp-pinecone"
]
}
}
Published Servers Configuration
"mcpServers": {
"mcp-pinecone": {
"command": "uvx",
"args": [
"--index-name",
"{your-index-name}",
"--api-key",
"{your-secret-api-key}",
"mcp-pinecone"
]
}
}
You can sign up for a Pinecone account here.
Create a new index in Pinecone, replacing {your-index-name}
and get an API key from the Pinecone dashboard, replacing {your-secret-api-key}
in the config.
To prepare the package for distribution:
uv sync
uv build
This will create source and wheel distributions in the dist/
directory.
uv publish
Note: You'll need to set PyPI credentials via environment variables or command flags:
--token
or UV_PUBLISH_TOKEN
--username
/UV_PUBLISH_USERNAME
and --password
/UV_PUBLISH_PASSWORD
Since MCP servers run over stdio, debugging can be challenging. For the best debugging experience, we strongly recommend using the MCP Inspector.
You can launch the MCP Inspector via npm
with this command:
npx @modelcontextprotocol/inspector uv --directory {project_dir} run mcp-pinecone
Upon launching, the Inspector will display a URL that you can access in your browser to begin debugging.
This project is licensed under the MIT License. See the LICENSE file for details.
The source code is available on GitHub.
Send your ideas and feedback to me on Bluesky or by opening an issue.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "mcp-pinecone": { "command": "uvx", "args": [ "--index-name", "{your-index-name}", "--api-key", "{your-secret-api-key}", "mcp-pinecone" ] } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by modelcontextprotocol
A basic implementation of persistent memory using a local knowledge graph. This lets Claude remember information about the user across chats.
by topoteretes
Provides dynamic memory for AI agents through modular ECL (Extract, Cognify, Load) pipelines, enabling seamless integration with graph and vector stores using minimal code.
by basicmachines-co
Enables persistent, local‑first knowledge management by allowing LLMs to read and write Markdown files during natural conversations, building a traversable knowledge graph that stays under the user’s control.
by smithery-ai
Provides read and search capabilities for Markdown notes in an Obsidian vault for Claude Desktop and other MCP clients.
by chatmcp
Summarize chat messages by querying a local chat database and returning concise overviews.
by dmayboroda
Provides on‑premises conversational retrieval‑augmented generation (RAG) with configurable Docker containers, supporting fully local execution, ChatGPT‑based custom GPTs, and Anthropic Claude integration.
by GreatScottyMac
Provides a project‑specific memory bank that stores decisions, progress, architecture, and custom data, exposing a structured knowledge graph via MCP for AI assistants and IDE tools.
by andrea9293
Provides document management and AI-powered semantic search for storing, retrieving, and querying text, markdown, and PDF files locally without external databases.
by scorzeth
Provides a local MCP server that interfaces with a running Anki instance to retrieve, create, and update flashcards through standard MCP calls.