by cuongtl1992
Provides a bridge between LLM applications and the Unleash feature flag system, enabling AI apps to check, create, update, and list feature flags via the Model Context Protocol.
Enables LLM‑driven applications to interact with an Unleash feature‑flag backend. Through MCP calls, AI agents can query flag status, expose flag metadata, create new flags, update existing ones, and retrieve a list of projects.
npm i
npm run build
npm start
{
"mcpServers": {
"unleash": {
"command": "npx",
"args": ["-y", "unleash-mcp"],
"env": {
"UNLEASH_URL": "YOUR_UNLEASH_END_POINT",
"UNLEASH_API_TOKEN": "YOUR_UNLEASH_API_TOKEN",
"MCP_TRANSPORT": "stdio",
"MCP_HTTP_PORT": 3001
}
}
}
}
Q: Which Node.js version is required? A: Node.js v18 or higher.
Q: Do I need a TypeScript compiler installed globally?
A: No. The project’s npm run build
script uses the local dev dependencies.
Q: Can I use the server over HTTP instead of STDIO?
A: Yes. Set MCP_TRANSPORT
to http
(default) and optionally adjust MCP_HTTP_PORT
.
Q: How are environment variables passed to the server?
A: Through the env
object in the MCP server configuration (e.g., UNLEASH_URL
, UNLEASH_API_TOKEN
).
Q: Is there a way to test the server locally?
A: Run npm test
after building; the test suite covers flag retrieval and project listing.
A Model Context Protocol (MCP) server implementation that integrates with Unleash Feature Toggle system.
This project provides a bridge between LLM applications and Unleash feature flag system using the Model Context Protocol (MCP). It allows AI applications to:
# Install dependencies
npm i
The Unleash MCP Server acts as a bridge between LLM applications and the Unleash feature flag system:
+----------------+ +-------------------+ +----------------+
| | | | | |
| LLM App | <--> | Unleash MCP | <--> | Unleash API |
| (MCP Client) | | Server | | Server |
| | | | | |
+----------------+ +-------------------+ +----------------+
unleash-mcp-server/
├── src/
│ ├── index.ts # Main entry point
│ ├── server.ts # Server implementation
│ ├── config.ts # Configuration handling
│ ├── transport/ # MCP transport implementations
│ │ ├── http.ts # HTTP/SSE transport
│ │ └── stdio.ts # STDIO transport
│ ├── unleash/ # Unleash API client implementations
│ │ ├── unleash-client.ts # Main Unleash client
│ │ ├── get-feature-flag.ts
│ │ └── get-all-projects.ts
│ ├── resources/ # MCP resource implementations
│ │ ├── flags.ts # Feature flag resources
│ │ └── projects.ts # Project resources
│ ├── tools/ # MCP tool implementations
│ │ ├── get-flag.ts # Get feature flag tool
│ │ └── get-projects.ts # Get projects tool
│ └── prompts/ # MCP prompt implementations
│ ├── flag-check.ts # Check single flag
│ └── batch-flag-check.ts # Check multiple flags
├── tests/ # Tests
└── package.json # Project configuration
Naming Conventions:
feature-flag.ts
)UnleashClient
)getFlagStatus
)FeatureFlagConfig
)Imports:
Documentation:
# Compile TypeScript
npm run build
# Run the server
npm start
# Run tests
npm test
# MCP stdio inspect
npm run build
npx @modelcontextprotocol/inspector node dist/index.js
# MCP sse inspect
npm start
npx @modelcontextprotocol/inspector
For Claude or Cursor config:
{
"mcpServers": {
"unleash": {
"command": "npx",
"args": [
"-y",
"unleash-mcp"
],
"env": {
"UNLEASH_URL": "YOUR_UNLEASH_END_POINT",
"UNLEASH_API_TOKEN": "YOUR_UNLEASH_API_TOKEN",
"MCP_TRANSPORT": "stdio",
"MCP_HTTP_PORT": 3001
}
}
}
}
Contributions are welcome! Please feel free to submit a Pull Request.
If you find this project helpful, consider buying me a coffee!
Scan the QR code above or click here to support the development of this project.
This project is licensed under the MIT License - see the LICENSE file for details.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "unleash": { "command": "npx", "args": [ "-y", "unleash-mcp" ], "env": { "UNLEASH_URL": "YOUR_UNLEASH_END_POINT", "UNLEASH_API_TOKEN": "YOUR_UNLEASH_API_TOKEN", "MCP_TRANSPORT": "stdio", "MCP_HTTP_PORT": 3001 } } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by zed-industries
A high‑performance, multiplayer code editor designed for speed and collaboration.
by modelcontextprotocol
Model Context Protocol Servers
by modelcontextprotocol
A Model Context Protocol server for Git repository interaction and automation.
by modelcontextprotocol
A Model Context Protocol server that provides time and timezone conversion capabilities.
by cline
An autonomous coding assistant that can create and edit files, execute terminal commands, and interact with a browser directly from your IDE, operating step‑by‑step with explicit user permission.
by continuedev
Enables faster shipping of code by integrating continuous AI agents across IDEs, terminals, and CI pipelines, offering chat, edit, autocomplete, and customizable agent workflows.
by upstash
Provides up-to-date, version‑specific library documentation and code examples directly inside LLM prompts, eliminating outdated information and hallucinated APIs.
by github
Connects AI tools directly to GitHub, enabling natural‑language interactions for repository browsing, issue and pull‑request management, CI/CD monitoring, code‑security analysis, and team collaboration.
by daytonaio
Provides a secure, elastic infrastructure that creates isolated sandboxes for running AI‑generated code with sub‑90 ms startup, unlimited persistence, and OCI/Docker compatibility.