by szeider
Enables AI agents to query large‑context language models with collections of files that exceed the agent’s own context window.
Consult7 provides a Model Context Protocol (MCP) server that lets AI agents send a set of files (paths with optional wildcards) together with a natural‑language query to a language model that offers a massive context window. The model processes the full assembled context and returns the answer directly to the agent.
uvx downloads and runs the tool in an isolated environment.# OpenRouter
claude mcp add -s user consult7 uvx -- consult7 openrouter <your-api-key>
# Google AI
claude mcp add -s user consult7 uvx -- consult7 google <your-api-key>
# OpenAI
claude mcp add -s user consult7 uvx -- consult7 openai <your-api-key>
For Claude Desktop edit the mcpServers section of the config file accordingly.gemini-2.5-flash|thinking) and a list of absolute file paths/wildcards.uvx consult7 <provider> <api-key> --test.uvx handles isolated execution.Q: Do I need to install anything locally?
A: No. uvx automatically downloads the Consult7 binary in a sandboxed environment.
Q: Which models can I use?
A: Provider‑specific examples are listed in the README – e.g., gemini-2.5-pro for Google, anthropic/claude-opus-4.1|thinking for OpenRouter, gpt-5|400k for OpenAI.
Q: How are large files handled? A: Files larger than 1 MB are ignored; the total payload is capped at 4 MB to stay within typical context limits.
Q: Can I customize reasoning tokens?
A: Yes, using the |thinking=30000 suffix, though the default is usually sufficient.
Q: How do I remove the server?
A: Run claude mcp remove consult7 -s user.
Consult7 is a Model Context Protocol (MCP) server that enables AI agents to consult large context window models via OpenRouter for analyzing extensive file collections - entire codebases, document repositories, or mixed content that exceed the current agent's context limits.
Consult7 enables any MCP-compatible agent to offload file analysis to large context models (up to 2M tokens). Useful when:
"For Claude Code users, Consult7 is a game changer."
Consult7 collects files from the specific paths you provide (with optional wildcards in filenames), assembles them into a single context, and sends them to a large context window model along with your query. The result is directly fed back to the agent you are working with.
["/Users/john/project/src/*.py", "/Users/john/project/lib/*.py"]"google/gemini-2.5-flash""fast"["/Users/john/webapp/src/*.py", "/Users/john/webapp/auth/*.py", "/Users/john/webapp/api/*.js"]"anthropic/claude-sonnet-4.5""think"["/Users/john/project/src/*.py", "/Users/john/project/tests/*.py"]"google/gemini-2.5-pro""think""/Users/john/reports/code_review.md""Result has been saved to /Users/john/reports/code_review.md" instead of flooding the agent's contextConsult7 now supports Google's Gemini 3 Pro (google/gemini-3-pro-preview) - the flagship reasoning model with a 1M context window and state-of-the-art performance on reasoning benchmarks.
Quick mnemonics for power users:
gemt = Gemini 3 Pro + think mode (flagship reasoning)gptt = GPT-5.1 + think mode (latest GPT)grot = Grok 4 + think mode (alternative reasoning)gemf = Gemini Flash Lite + fast mode (ultra fast)ULTRA = Run GEMT, GPTT, and GROT in parallel for maximum insightThese mnemonics make it easy to reference model+mode combinations in your queries.
Simply run:
claude mcp add -s user consult7 uvx -- consult7 your-openrouter-api-key
Add to your Claude Desktop configuration file:
{
"mcpServers": {
"consult7": {
"type": "stdio",
"command": "uvx",
"args": ["consult7", "your-openrouter-api-key"]
}
}
}
Replace your-openrouter-api-key with your actual OpenRouter API key.
No installation required - uvx automatically downloads and runs consult7 in an isolated environment.
uvx consult7 <api-key> [--test]
<api-key>: Required. Your OpenRouter API key--test: Optional. Test the API connectionThe model and mode are specified when calling the tool, not at startup.
Consult7 supports all 500+ models available on OpenRouter. Below are the flagship models with optimized dynamic file size limits:
| Model | Context | Use Case |
|---|---|---|
openai/gpt-5.1 |
400k | Latest GPT, balanced performance |
google/gemini-3-pro-preview |
1M | Flagship reasoning model |
google/gemini-2.5-pro |
1M | Best for complex analysis |
google/gemini-2.5-flash |
1M | Fast, good for most tasks |
google/gemini-2.5-flash-lite |
1M | Ultra fast, simple queries |
anthropic/claude-sonnet-4.5 |
1M | Excellent reasoning |
anthropic/claude-opus-4.5 |
200k | Best quality, slower |
x-ai/grok-4 |
256k | Alternative reasoning model |
x-ai/grok-4-fast |
2M | Largest context window |
Quick mnemonics:
gptt = openai/gpt-5.1 + think (latest GPT, deep reasoning)gemt = google/gemini-3-pro-preview + think (Gemini 3 Pro, flagship reasoning)grot = x-ai/grok-4 + think (Grok 4, deep reasoning)oput = anthropic/claude-opus-4.5 + think (Claude Opus, deep reasoning)opuf = anthropic/claude-opus-4.5 + fast (Claude Opus, no reasoning)gemf = google/gemini-2.5-flash-lite + fast (ultra fast)ULTRA = call GEMT, GPTT, GROT, and OPUT IN PARALLEL (4 frontier models for maximum insight)You can use any OpenRouter model ID (e.g., deepseek/deepseek-r1-0528). See the full model list. File size limits are automatically calculated based on each model's context window.
fast: No reasoning - quick answers, simple tasksmid: Moderate reasoning - code reviews, bug analysisthink: Maximum reasoning - security audits, complex refactoring/Users/john/project/src/*.py/Users/john/project/*.py (not in directory paths)*.py not *["/path/src/*.py", "/path/README.md", "/path/tests/*_test.py"]Common patterns:
/path/to/dir/*.py/path/to/tests/*_test.py or /path/to/tests/test_*.py["/path/*.js", "/path/*.ts"]Automatically ignored: __pycache__, .env, secrets.py, .DS_Store, .git, node_modules
Size limits: Dynamic based on model context window (e.g., Grok 4 Fast: ~8MB, GPT-5.1: ~1.5MB)
The consultation tool accepts the following parameters:
fast, mid, or think_updated suffix (e.g., report.md → report_updated.md)"Result has been saved to /path/to/file"Claude Code will automatically use the tool with proper parameters:
{
"files": ["/Users/john/project/src/*.py"],
"query": "Explain the main architecture",
"model": "google/gemini-2.5-flash",
"mode": "mid"
}
from consult7.consultation import consultation_impl
result = await consultation_impl(
files=["/path/to/file.py"],
query="Explain this code",
model="google/gemini-2.5-flash",
mode="mid", # fast, mid, or think
provider="openrouter",
api_key="sk-or-v1-..."
)
# Test OpenRouter connection
uvx consult7 sk-or-v1-your-api-key --test
To remove consult7 from Claude Code:
claude mcp remove consult7 -s user
google/gemini-3-pro-preview (1M context, flagship reasoning model)gemt (Gemini 3 Pro), grot (Grok 4), ULTRA (parallel execution){"reasoning": {"enabled": true}} API format for Gemini 3 Pro|thinking suffix - use mode parameter instead (now required)mode parameter API: fast, mid, thinkconsult7 <provider> <key> to consult7 <key>output_file parameter to save responses to filesMIT
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by daytonaio
Provides a secure, elastic infrastructure that creates isolated sandboxes for running AI‑generated code with sub‑90 ms startup, unlimited persistence, and OCI/Docker compatibility.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by github
Connects AI tools directly to GitHub, enabling natural‑language interactions for repository browsing, issue and pull‑request management, CI/CD monitoring, code‑security analysis, and team collaboration.