by metoro-io
Provides an MCP server that exposes Metoro's eBPF‑based telemetry APIs to large language models, enabling AI‑driven queries and insights about Kubernetes clusters.
The server acts as a bridge between the Claude Desktop App (or any LLM client supporting MCP) and Metoro's observability platform. By running the server, LLMs can retrieve real‑time telemetry, service topology, and performance data from a Kubernetes cluster without writing custom code.
go build -o metoro-mcp-server.~/Library/Application Support/Claude/claude_desktop_config.json to point to the built binary and supply environment variables:
{
"mcpServers": {
"metoro-mcp-server": {
"command": "<path>/metoro-mcp-server",
"args": [],
"env": {
"METORO_AUTH_TOKEN": "<your token>",
"METORO_API_URL": "https://us-east.metoro.io"
}
}
}
}
Q: Do I need a Metoro account? A: No. You can use the public demo token and API URL to experiment.
Q: Which platforms are supported? A: Any system that can run Go binaries and the Claude Desktop App (macOS, Linux, Windows via WSL).
Q: How is authentication handled?
A: Via a JWT token passed in the METORO_AUTH_TOKEN environment variable.
Q: Can I customize the server behavior? A: The server is built on the Golang MCP SDK; extending it requires modifying the source and rebuilding.
Q: What if I want to use a different LLM client? A: Any client that implements the Model Context Protocol can communicate with this server by following the same configuration pattern.
This repository contains th Metoro MCP (Model Context Protocol) Server. This MCP Server allows you to interact with your Kubernetes cluster via the Claude Desktop App!
You can read more about the Model Context Protocol here: https://modelcontextprotocol.io
But in a nutshell
The Model Context Protocol (MCP) is an open protocol that enables seamless integration between LLM applications and external data sources and tools. Whether you’re building an AI-powered IDE, enhancing a chat interface, or creating custom AI workflows, MCP provides a standardized way to connect LLMs with the context they need.
Metoro is an observability platform designed for microservices running in Kubernetes and uses eBPF based instrumentation to generate deep telemetry without code changes. The data that is generated by the eBPF agents is sent to Metoro's backend to be stored and in the Metoro frontend using our apis.
This MCP server exposes those APIs to an LLM so you can ask your AI questions about your Kubernetes cluster.
https://github.com/user-attachments/assets/b3f21e9a-45b8-4c17-8d8c-cff560d8694f
brew install go for mac or sudo apt-get install golang for ubuntu.git clone https://github.com/metoro-io/metoro-mcp-server.gitcd metoro-mcp-servergo build -o metoro-mcp-serverCopy your auth token from your Metoro account in Settings -> Users Settings.
Create a file in ~/Library/Application Support/Claude/claude_desktop_config.json with the following contents:
{
"mcpServers": {
"metoro-mcp-server": {
"command": "<your path to Metoro MCP server go executable>/metoro-mcp-server",
"args": [],
"env": {
"METORO_AUTH_TOKEN" : "<your auth token>",
"METORO_API_URL": "https://us-east.metoro.io"
}
}
}
}
No worries, you can still play around using the Live Demo Cluster.
The included token is a demo token, publicly available for anyone to use.
Create a file in ~/Library/Application Support/Claude/claude_desktop_config.json with the following contents:
{
"mcpServers": {
"metoro-mcp-server": {
"command": "<your path to Metoro MCP server go executable>/metoro-mcp-server",
"args": [],
"env": {
"METORO_AUTH_TOKEN" : "eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJjdXN0b21lcklkIjoiOThlZDU1M2QtYzY4ZC00MDRhLWFhZjItNDM2ODllNWJiMGUzIiwiZW1haWwiOiJ0ZXN0QGNocmlzYmF0dGFyYmVlLmNvbSIsImV4cCI6MTgyMTI0NzIzN30.7G6alDpcZh_OThYj293Jce5rjeOBqAhOlANR_Fl5auw",
"METORO_API_URL": "https://demo.us-east.metoro.io"
}
}
}
}
claude_desktop_config.json save the file and restart Claude Desktop app.This server is built on top of our Golang MCP SDK.
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by netdata
Delivers real‑time, per‑second infrastructure monitoring with zero‑configuration agents, on‑edge machine‑learning anomaly detection, and built‑in dashboards.
by Arize-ai
Open-source AI observability platform enabling tracing, evaluation, dataset versioning, experiment tracking, prompt management, and interactive playground for LLM applications.
by msgbyte
Provides integrated website traffic analysis, uptime checking, and server health monitoring in a single self‑hosted platform.
by grafana
Provides programmatic access to a Grafana instance and its surrounding ecosystem through the Model Context Protocol, enabling AI assistants and other clients to query and manipulate dashboards, datasources, alerts, incidents, on‑call schedules, and more.
by dynatrace-oss
Provides a local server that enables real‑time interaction with the Dynatrace observability platform, exposing tools for querying data, retrieving problems, sending Slack notifications, and integrating AI assistance.
by pydantic
Provides tools to retrieve and query OpenTelemetry trace and metric data from Pydantic Logfire, allowing LLMs to analyze distributed traces and run arbitrary SQL queries against telemetry records.
by VictoriaMetrics-Community
Provides a Model Context Protocol server exposing read‑only VictoriaMetrics APIs, enabling seamless monitoring, observability, and automation through AI‑driven assistants.
by GeLi2001
Enables interaction with the Datadog API through a Model Context Protocol server, providing access to monitors, dashboards, metrics, logs, events, and incident data.
by last9
Provides AI agents with real‑time production context—including logs, metrics, traces, and alerts—through a Model Context Protocol server, enabling automatic code fixing and faster debugging.