by szeider
Enables AI agents to query large‑context language models with collections of files that exceed the agent’s own context window.
Consult7 provides a Model Context Protocol (MCP) server that lets AI agents send a set of files (paths with optional wildcards) together with a natural‑language query to a language model that offers a massive context window. The model processes the full assembled context and returns the answer directly to the agent.
uvx
downloads and runs the tool in an isolated environment.# OpenRouter
claude mcp add -s user consult7 uvx -- consult7 openrouter <your-api-key>
# Google AI
claude mcp add -s user consult7 uvx -- consult7 google <your-api-key>
# OpenAI
claude mcp add -s user consult7 uvx -- consult7 openai <your-api-key>
For Claude Desktop edit the mcpServers
section of the config file accordingly.gemini-2.5-flash|thinking
) and a list of absolute file paths/wildcards.uvx consult7 <provider> <api-key> --test
.uvx
handles isolated execution.Q: Do I need to install anything locally?
A: No. uvx
automatically downloads the Consult7 binary in a sandboxed environment.
Q: Which models can I use?
A: Provider‑specific examples are listed in the README – e.g., gemini-2.5-pro
for Google, anthropic/claude-opus-4.1|thinking
for OpenRouter, gpt-5|400k
for OpenAI.
Q: How are large files handled? A: Files larger than 1 MB are ignored; the total payload is capped at 4 MB to stay within typical context limits.
Q: Can I customize reasoning tokens?
A: Yes, using the |thinking=30000
suffix, though the default is usually sufficient.
Q: How do I remove the server?
A: Run claude mcp remove consult7 -s user
.
Consult7 is a Model Context Protocol (MCP) server that enables AI agents to consult large context window models for analyzing extensive file collections - entire codebases, document repositories, or mixed content that exceed the current agent's context limits. Supports providers Openrouter, OpenAI, and Google.
When working with AI agents that have limited context windows (like Claude with 200K tokens), Consult7 allows them to leverage models with massive context windows to analyze large codebases or document collections that would otherwise be impossible to process in a single query.
"For Claude Code users, Consult7 is a game changer."
Consult7 collects files from the specific paths you provide (with optional wildcards in filenames), assembles them into a single context, and sends them to a large context window model along with your query. The result is directly fed back to the agent you are working with.
["/Users/john/project/src/*.py", "/Users/john/project/lib/*.py"]
"gemini-2.5-flash"
["/Users/john/backend/src/*.py", "/Users/john/backend/auth/*.js"]
"gemini-2.5-pro"
["/Users/john/project/tests/*_test.py", "/Users/john/project/src/*.py"]
"gemini-2.5-flash"
["/Users/john/webapp/src/*.py", "/Users/john/webapp/auth/*.py", "/Users/john/webapp/api/*.js"]
"gemini-2.5-flash|thinking"
Simply run:
# OpenRouter
claude mcp add -s user consult7 uvx -- consult7 openrouter your-api-key
# Google AI
claude mcp add -s user consult7 uvx -- consult7 google your-api-key
# OpenAI
claude mcp add -s user consult7 uvx -- consult7 openai your-api-key
Add to your Claude Desktop configuration file:
{
"mcpServers": {
"consult7": {
"type": "stdio",
"command": "uvx",
"args": ["consult7", "openrouter", "your-api-key"]
}
}
}
Replace openrouter
with your provider choice (google
or openai
) and your-api-key
with your actual API key.
No installation required - uvx
automatically downloads and runs consult7 in an isolated environment.
uvx consult7 <provider> <api-key> [--test]
<provider>
: Required. Choose from openrouter
, google
, or openai
<api-key>
: Required. Your API key for the chosen provider--test
: Optional. Test the API connectionThe model is specified when calling the tool, not at startup. The server shows example models for your provider on startup.
Standard models:
"gemini-2.5-flash"
- Fast model"gemini-2.5-flash-lite"
- Ultra fast lite model"gemini-2.5-pro"
- Intelligent model"gemini-2.0-flash-exp"
- Experimental modelWith thinking mode (add |thinking
suffix):
"gemini-2.5-flash|thinking"
- Fast with deep reasoning"gemini-2.5-flash-lite|thinking"
- Ultra fast with deep reasoning"gemini-2.5-pro|thinking"
- Intelligent with deep reasoningStandard models:
"google/gemini-2.5-pro"
- Intelligent, 1M context"google/gemini-2.5-flash"
- Fast, 1M context"google/gemini-2.5-flash-lite"
- Ultra fast, 1M context"anthropic/claude-sonnet-4"
- Claude Sonnet, 200k context"anthropic/claude-opus-4.1"
- Claude Opus 4.1, 200k context"openai/gpt-5"
- GPT-5, 400k context"openai/gpt-4.1"
- GPT-4.1, 1M+ contextWith reasoning mode (add |thinking
suffix):
"anthropic/claude-sonnet-4|thinking"
- Claude with 31,999 reasoning tokens"anthropic/claude-opus-4.1|thinking"
- Opus 4.1 with reasoning"google/gemini-2.5-flash-lite|thinking"
- Ultra fast with reasoning"openai/gpt-5|thinking"
- GPT-5 with reasoning"openai/gpt-4.1|thinking"
- GPT-4.1 with reasoning effort=highStandard models (include context length):
"gpt-5|400k"
- GPT-5, 400k context"gpt-5-mini|400k"
- GPT-5 Mini, faster"gpt-5-nano|400k"
- GPT-5 Nano, ultra fast"gpt-4.1-2025-04-14|1047576"
- 1M+ context, very fast"gpt-4.1-nano-2025-04-14|1047576"
- 1M+ context, ultra fast"o3-2025-04-16|200k"
- Advanced reasoning model"o4-mini-2025-04-16|200k"
- Fast reasoning modelO-series models with |thinking marker:
"o1-mini|128k|thinking"
- Mini reasoning with |thinking marker"o3-2025-04-16|200k|thinking"
- Advanced reasoning with |thinking markerNote: For OpenAI, |thinking is only supported on o-series models and serves as an informational marker. The models use reasoning tokens automatically.
Advanced: You can specify custom thinking tokens with |thinking=30000
but this is rarely needed.
When using the consultation tool, you provide a list of file paths with these rules:
All paths must be absolute (start with /
)
/Users/john/project/src/*.py
src/*.py
or ./src/*.py
Wildcards (*
) only allowed in filenames, not in directory paths
/Users/john/project/*.py
/Users/*/project/*.py
or /Users/john/**/*.py
Must specify extension when using wildcards
/Users/john/project/*.py
/Users/john/project/*
Mix specific files and patterns freely
["/path/src/*.py", "/path/README.md", "/path/tests/*_test.py"]
Common patterns:
/path/to/dir/*.py
/path/to/tests/*_test.py
or /path/to/tests/test_*.py
["/path/*.js", "/path/*.ts"]
The tool automatically ignores: __pycache__
, .env
, secrets.py
, .DS_Store
, .git
, node_modules
Size limits: 1MB per file, 4MB total (optimized for ~1M token context windows)
# Test OpenRouter
uvx consult7 openrouter sk-or-v1-... --test
# Test Google AI
uvx consult7 google AIza... --test
# Test OpenAI
uvx consult7 openai sk-proj-... --test
To remove consult7 from Claude Code (or before reinstalling):
claude mcp remove consult7 -s user
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by github
Connects AI tools directly to GitHub, enabling natural‑language interactions for repository browsing, issue and pull‑request management, CI/CD monitoring, code‑security analysis, and team collaboration.
by daytonaio
Provides a secure, elastic infrastructure that creates isolated sandboxes for running AI‑generated code with sub‑90 ms startup, unlimited persistence, and OCI/Docker compatibility.