by axiomhq
Enables AI agents to query Axiom data using Axiom Processing Language (APL) through a set of Model Context Protocol tools.
Provides a Model Context Protocol (MCP) server implementation that connects AI agents (e.g., Claude) to Axiom's data platform, allowing execution of APL queries, dataset discovery, schema retrieval, saved query access, and monitor management.
go install github.com/axiomhq/axiom-mcp@latest
.AXIOM_TOKEN
, AXIOM_URL
).Q: Which authentication method is required?
A: A valid Axiom API token must be supplied via the token
parameter, config file, flag, or AXIOM_TOKEN
environment variable.
Q: Can this server handle MCP resources or prompts? A: Currently only the six listed tools are supported; resources and prompts are not yet implemented.
Q: How are rate limits configured?
A: Use the *-rate
and *-burst
options for queries, datasets, and monitors, either in the config file, flags, or corresponding environment variables.
Q: Is there a Docker image available? A: The README does not mention a Docker image; the recommended approach is to use the binary directly.
A Model Context Protocol server implementation for Axiom that enables AI agents to query your data using Axiom Processing Language (APL).
Works with Claude desktop app. Implements six MCP tools:
Note: All tools require an API token for authentication. Use your API token as the token
parameter.
No support for MCP resources or prompts yet.
Download the latest built binary from the releases page.
go install github.com/axiomhq/axiom-mcp@latest
Configure using one of these methods:
token xaat-your-api-token
url https://api.axiom.co
query-rate 1
query-burst 1
datasets-rate 1
datasets-burst 1
monitors-rate 1
monitors-burst 1
axiom-mcp \
-token xaat-your-api-token \
-url https://api.axiom.co \
-query-rate 1 \
-query-burst 1 \
-datasets-rate 1 \
-datasets-burst 1 \
-monitors-rate 1 \
-monitors-burst 1
export AXIOM_TOKEN=xaat-your-api-token
export AXIOM_URL=https://api.axiom.co
export AXIOM_QUERY_RATE=1
export AXIOM_QUERY_BURST=1
export AXIOM_DATASETS_RATE=1
export AXIOM_DATASETS_BURST=1
export AXIOM_MONITORS_RATE=1
export AXIOM_MONITORS_BURST=1
echo "token xaat-your-api-token" > config.txt
code ~/Library/Application\ Support/Claude/claude_desktop_config.json
{
"mcpServers": {
"axiom": {
"command": "/path/to/your/axiom-mcp-binary",
"args" : ["--config", "/path/to/your/config.txt"],
"env": { // Alternatively, you can set the environment variables here
"AXIOM_TOKEN": "xaat-your-api-token",
"AXIOM_URL": "https://api.axiom.co"
}
}
}
}
MIT License - see LICENSE file
Please log in to share your review and rating for this MCP.
Explore related MCPs that share similar capabilities and solve comparable challenges
by modelcontextprotocol
An MCP server implementation that provides a tool for dynamic and reflective problem-solving through a structured thinking process.
by danny-avila
Provides a self‑hosted ChatGPT‑style interface supporting numerous AI models, agents, code interpreter, image generation, multimodal interactions, and secure multi‑user authentication.
by block
Automates engineering tasks on local machines, executing code, building projects, debugging, orchestrating workflows, and interacting with external APIs using any LLM.
by RooCodeInc
Provides an autonomous AI coding partner inside the editor that can understand natural language, manipulate files, run commands, browse the web, and be customized via modes and instructions.
by pydantic
A Python framework that enables seamless integration of Pydantic validation with large language models, providing type‑safe agent construction, dependency injection, and structured output handling.
by lastmile-ai
Build effective agents using Model Context Protocol and simple, composable workflow patterns.
by mcp-use
A Python SDK that simplifies interaction with MCP servers and enables developers to create custom agents with tool‑calling capabilities.
by nanbingxyz
A cross‑platform desktop AI assistant that connects to major LLM providers, supports a local knowledge base, and enables tool integration via MCP servers.
by gptme
Provides a personal AI assistant that runs directly in the terminal, capable of executing code, manipulating files, browsing the web, using vision, and interfacing with various LLM providers.