by YuChenSSR
Queries multiple Ollama models, combines their perspectives, and synthesizes a comprehensive answer for a single question, enabling a council‑of‑advisors style interaction.
Multi Model Advisor lets a user pose a single query and receive blended insights from several locally hosted Ollama models, each acting with a distinct persona. The server gathers each model's response and returns the collection so Claude (or any MCP‑compatible client) can produce a unified answer.
npx -y @smithery/cli install @YuChenSSR/multi-ai-advisor-mcp --client claude
or manually clone, run npm install
, and npm run build
..env
file with Ollama URL, default model list, and system prompts for each model.claude_desktop_config.json
pointing to the built index.js
(or use the npx‑based command in the serverConfig
below).what are the most important skills for success in today's job market, you can use gemma3:1b, llama3.2:1b, deepseek-r1:1.5b to help you
Claude will invoke the query-models
tool, the server contacts the three models, and Claude synthesizes the replies.list-available-models
tool to inspect installed Ollama models.Q: Do I need an internet connection? A: No. All models run locally via Ollama; only the local Ollama API endpoint is required.
Q: Which models can I use?
A: Any model installed in Ollama. The default set includes gemma3:1b
, llama3.2:1b
, and deepseek-r1:1.5b
, but you can modify DEFAULT_MODELS
in .env
.
Q: How do I change a model’s persona?
A: Edit the corresponding <MODEL>_SYSTEM_PROMPT
variable in .env
and restart the server.
Q: What if a model fails to respond? A: The server returns the responses it received; missing models are omitted, and Claude will synthesize from the available data.
Q: Can I add more than three models?
A: Yes. List any number of model names in DEFAULT_MODELS
(comma‑separated) and provide matching system prompts.
Q: How do I troubleshoot Ollama connection issues?
A: Verify Ollama is running (ollama serve
), confirm OLLAMA_API_URL
points to the correct host/port, and test the endpoint in a browser.
Q: Is there a memory limit? A: Larger models consume more RAM. If you encounter out‑of‑memory errors, switch to smaller models or increase system memory.
A Model Context Protocol (MCP) server that queries multiple Ollama models and combines their responses, providing diverse AI perspectives on a single question. This creates a "council of advisors" approach where Claude can synthesize multiple viewpoints alongside its own to provide more comprehensive answers.
To install multi-ai-advisor-mcp for Claude Desktop automatically via Smithery:
npx -y @smithery/cli install @YuChenSSR/multi-ai-advisor-mcp --client claude
Clone this repository:
git clone https://github.com/YuChenSSR/multi-ai-advisor-mcp.git
cd multi-ai-advisor-mcp
Install dependencies:
npm install
Build the project:
npm run build
Install required Ollama models:
ollama pull gemma3:1b
ollama pull llama3.2:1b
ollama pull deepseek-r1:1.5b
Create a .env
file in the project root with your desired configuration:
# Server configuration
SERVER_NAME=multi-model-advisor
SERVER_VERSION=1.0.0
DEBUG=true
# Ollama configuration
OLLAMA_API_URL=http://localhost:11434
DEFAULT_MODELS=gemma3:1b,llama3.2:1b,deepseek-r1:1.5b
# System prompts for each model
GEMMA_SYSTEM_PROMPT=You are a creative and innovative AI assistant. Think outside the box and offer novel perspectives.
LLAMA_SYSTEM_PROMPT=You are a supportive and empathetic AI assistant focused on human well-being. Provide considerate and balanced advice.
DEEPSEEK_SYSTEM_PROMPT=You are a logical and analytical AI assistant. Think step-by-step and explain your reasoning clearly.
Locate your Claude for Desktop configuration file:
~/Library/Application Support/Claude/claude_desktop_config.json
%APPDATA%\Claude\claude_desktop_config.json
Edit the file to add the Multi-Model Advisor MCP server:
{
"mcpServers": {
"multi-model-advisor": {
"command": "node",
"args": ["/absolute/path/to/multi-ai-advisor-mcp/build/index.js"]
}
}
}
Replace /absolute/path/to/
with the actual path to your project directory
Restart Claude for Desktop
Once connected to Claude for Desktop, you can use the Multi-Model Advisor in several ways:
You can see all available models on your system:
Show me which Ollama models are available on my system
This will display all installed Ollama models and indicate which ones are configured as defaults.
Simply ask Claude to use the multi-model advisor:
what are the most important skills for success in today's job market,
you can use gemma3:1b, llama3.2:1b, deepseek-r1:1.5b to help you
Claude will query all default models and provide a synthesized response based on their different perspectives.
The MCP server exposes two tools:
list-available-models
: Shows all Ollama models on your systemquery-models
: Queries multiple models with a questionWhen you ask Claude a question referring to the multi-model advisor:
query-models
toolEach model can have a different "persona" or role assigned, encouraging diverse perspectives.
If the server can't connect to Ollama:
ollama serve
)If a model is reported as unavailable:
ollama pull <model-name>
ollama list
list-available-models
tool to see all available modelsIf the tools don't appear in Claude:
Some managers' AI models may have chosen larger models, but there is not enough memory to run them. You can try specifying a smaller model (see the Basic Usage) or upgrading the memory.
MIT License
For more details, please see the LICENSE file in this project repository
Contributions are welcome! Please feel free to submit a Pull Request.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "multi-model-advisor": { "command": "npx", "args": [ "-y", "@YuChenSSR/multi-ai-advisor-mcp" ], "env": {} } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by modelcontextprotocol
An MCP server implementation that provides a tool for dynamic and reflective problem-solving through a structured thinking process.
by danny-avila
Provides a self‑hosted ChatGPT‑style interface supporting numerous AI models, agents, code interpreter, image generation, multimodal interactions, and secure multi‑user authentication.
by block
Automates engineering tasks on local machines, executing code, building projects, debugging, orchestrating workflows, and interacting with external APIs using any LLM.
by RooCodeInc
Provides an autonomous AI coding partner inside the editor that can understand natural language, manipulate files, run commands, browse the web, and be customized via modes and instructions.
by pydantic
A Python framework that enables seamless integration of Pydantic validation with large language models, providing type‑safe agent construction, dependency injection, and structured output handling.
by lastmile-ai
Build effective agents using Model Context Protocol and simple, composable workflow patterns.
by mcp-use
A Python SDK that simplifies interaction with MCP servers and enables developers to create custom agents with tool‑calling capabilities.
by nanbingxyz
A cross‑platform desktop AI assistant that connects to major LLM providers, supports a local knowledge base, and enables tool integration via MCP servers.
by gptme
Provides a personal AI assistant that runs directly in the terminal, capable of executing code, manipulating files, browsing the web, using vision, and interfacing with various LLM providers.