by vectorize-io
Provides a Model Context Protocol (MCP) server that integrates with Vectorize to enable vector‑based document retrieval, text extraction, and deep research capabilities for LLM workflows.
Enables MCP‑compatible services that connect to Vectorize’s retrieval and extraction APIs, allowing AI applications to search across vector databases, convert files to markdown, and generate private deep‑research reports.
export VECTORIZE_ORG_ID=YOUR_ORG_ID
export VECTORIZE_TOKEN=YOUR_TOKEN
export VECTORIZE_PIPELINE_ID=YOUR_PIPELINE_ID
npx -y @vectorize-io/vectorize-mcp-server@latest
retrieve
, extract
, deep‑research
) via MCP.k
results.npm install
+ npm run dev
for local testing.Q: Do I need to install anything globally?
A: No. The server runs via npx
, which fetches the latest package on demand.
Q: Which environment variables are required?
A: VECTORIZE_ORG_ID
, VECTORIZE_TOKEN
, and VECTORIZE_PIPELINE_ID
must be set before starting the server.
Q: Can I customize the server name in MCP configurations?
A: Yes. The README uses vectorize
as the server identifier, but any name can be used as long as the MCP client points to the correct command and args.
Q: What Node.js version is supported? A: The package follows the standard Node.js LTS releases; using the latest LTS (e.g., 20.x) is recommended.
Q: Is the source code open source? A: Yes, the repository is public under the MIT license on GitHub.
A Model Context Protocol (MCP) server implementation that integrates with Vectorize for advanced Vector retrieval and text extraction.
export VECTORIZE_ORG_ID=YOUR_ORG_ID
export VECTORIZE_TOKEN=YOUR_TOKEN
export VECTORIZE_PIPELINE_ID=YOUR_PIPELINE_ID
npx -y @vectorize-io/vectorize-mcp-server@latest
For one-click installation, click one of the install buttons below:
For the quickest installation, use the one-click install buttons at the top of this section.
To install manually, add the following JSON block to your User Settings (JSON) file in VS Code. You can do this by pressing Ctrl + Shift + P
and typing Preferences: Open User Settings (JSON)
.
{
"mcp": {
"inputs": [
{
"type": "promptString",
"id": "org_id",
"description": "Vectorize Organization ID"
},
{
"type": "promptString",
"id": "token",
"description": "Vectorize Token",
"password": true
},
{
"type": "promptString",
"id": "pipeline_id",
"description": "Vectorize Pipeline ID"
}
],
"servers": {
"vectorize": {
"command": "npx",
"args": ["-y", "@vectorize-io/vectorize-mcp-server@latest"],
"env": {
"VECTORIZE_ORG_ID": "${input:org_id}",
"VECTORIZE_TOKEN": "${input:token}",
"VECTORIZE_PIPELINE_ID": "${input:pipeline_id}"
}
}
}
}
}
Optionally, you can add the following to a file called .vscode/mcp.json
in your workspace to share the configuration with others:
{
"inputs": [
{
"type": "promptString",
"id": "org_id",
"description": "Vectorize Organization ID"
},
{
"type": "promptString",
"id": "token",
"description": "Vectorize Token",
"password": true
},
{
"type": "promptString",
"id": "pipeline_id",
"description": "Vectorize Pipeline ID"
}
],
"servers": {
"vectorize": {
"command": "npx",
"args": ["-y", "@vectorize-io/vectorize-mcp-server@latest"],
"env": {
"VECTORIZE_ORG_ID": "${input:org_id}",
"VECTORIZE_TOKEN": "${input:token}",
"VECTORIZE_PIPELINE_ID": "${input:pipeline_id}"
}
}
}
}
{
"mcpServers": {
"vectorize": {
"command": "npx",
"args": ["-y", "@vectorize-io/vectorize-mcp-server@latest"],
"env": {
"VECTORIZE_ORG_ID": "your-org-id",
"VECTORIZE_TOKEN": "your-token",
"VECTORIZE_PIPELINE_ID": "your-pipeline-id"
}
}
}
}
Perform vector search and retrieve documents (see official API):
{
"name": "retrieve",
"arguments": {
"question": "Financial health of the company",
"k": 5
}
}
Extract text from a document and chunk it into Markdown format (see official API):
{
"name": "extract",
"arguments": {
"base64document": "base64-encoded-document",
"contentType": "application/pdf"
}
}
Generate a Private Deep Research from your pipeline (see official API):
{
"name": "deep-research",
"arguments": {
"query": "Generate a financial status report about the company",
"webSearch": true
}
}
npm install
npm run dev
Change the package.json version and then:
git commit -am "x.y.z"
git tag x.y.z
git push origin
git push origin --tags
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "vectorize": { "command": "npx", "args": [ "-y", "@vectorize-io/vectorize-mcp-server@latest" ], "env": { "VECTORIZE_ORG_ID": "<YOUR_ORG_ID>", "VECTORIZE_TOKEN": "<YOUR_TOKEN>", "VECTORIZE_PIPELINE_ID": "<YOUR_PIPELINE_ID>" } } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by exa-labs
Provides real-time web search capabilities to AI assistants via a Model Context Protocol server, enabling safe and controlled access to the Exa AI Search API.
by elastic
Enables natural‑language interaction with Elasticsearch indices via the Model Context Protocol, exposing tools for listing indices, fetching mappings, performing searches, running ES|QL queries, and retrieving shard information.
by graphlit
Enables integration between MCP clients and the Graphlit platform, providing ingestion, extraction, retrieval, and RAG capabilities across a wide range of data sources and connectors.
by mamertofabian
Fast cross‑platform file searching leveraging the Everything SDK on Windows, Spotlight on macOS, and locate/plocate on Linux.
by cr7258
Provides Elasticsearch and OpenSearch interaction via Model Context Protocol, enabling document search, index management, cluster monitoring, and alias operations.
by liuyoshio
Provides natural‑language search and recommendation for Model Context Protocol servers, delivering rich metadata and real‑time updates.
by ihor-sokoliuk
Provides web search capabilities via the SearXNG API, exposing them through an MCP server for seamless integration with AI agents and tools.
by fatwang2
Provides web and news search, URL crawling, sitemap extraction, deep‑reasoning, and trending topic retrieval via Search1API, exposed as an MCP server for integration with AI clients.
by cnych
Provides SEO data retrieval via Ahrefs, exposing MCP tools for backlink analysis, keyword generation, traffic estimation, and keyword difficulty, with automated CAPTCHA solving and response caching.