by Arize-ai
Open-source AI observability platform enabling tracing, evaluation, dataset versioning, experiment tracking, prompt management, and interactive playground for LLM applications.
Provides end‑to‑end observability for large language model (LLM) workflows: captures runtime traces via OpenTelemetry, runs automated evaluations, stores versioned datasets, tracks experiments, and offers a UI playground for prompt engineering.
pip install arize-phoenix
Or pull the container: docker pull arizephoenix/phoenix.arize-phoenix serve).openinference-instrumentation-openai, openinference-instrumentation-langchain, etc.) or use the provided arize-phoenix-otel package for a simplified setup.http://localhost:8080 (default) to explore traces, evaluation results, datasets, and experiments.arize-phoenix-client (Python) or @arizeai/phoenix-client (JS/TS) to create datasets, upload evaluations, and query metadata.arize-phoenix-evals.Q: Do I need a cloud account to run Phoenix? A: No. Phoenix can run locally via Docker or the Python package. Cloud‑hosted instances are optional.
Q: Which frameworks are supported out of the box? A: LangChain, LlamaIndex, Haystack, DSPy, Smolagents, OpenAI SDK, Bedrock, Vertex AI, MistralAI, LiteLLM, and many more via OpenInference.
Q: How are evaluations stored? A: Evaluations are persisted as part of the Phoenix backend and can be queried through the UI or client SDKs.
Q: Can I extend the evaluation suite?
A: Yes. Custom eval functions can be added using arize-phoenix-evals or the TypeScript counterpart.
Q: Is there a licensing cost? A: Phoenix is released under the Elastic License 2.0 and is free to use; commercial support is offered by Arize AI.
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (🦙LlamaIndex, 🦜⛓LangChain, Haystack, 🧩DSPy, 🤗smolagents) and LLM providers (OpenAI, Bedrock, MistralAI, VertexAI, LiteLLM, Google GenAI and more). For details on auto-instrumentation, check out the OpenInference project.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Install Phoenix via pip or conda
pip install arize-phoenix
Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.
The arize-phoenix package includes the entire Phoenix platfom. However if you have deployed the Phoenix platform, there are light-weight Python sub-packages and TypeScript packages that can be used in conjunction with the platfrom.
| Package | Language | Description |
|---|---|---|
| arize-phoenix-otel | Python |
Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |
| arize-phoenix-client | Python |
Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface |
| arize-phoenix-evals | Python |
Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more |
| @arizeai/phoenix-client | JavaScript |
Client for the Arize Phoenix API |
| @arizeai/phoenix-evals | TypeScript |
TypeScript evaluation library for LLM applications (alpha release) |
| @arizeai/phoenix-mcp | JavaScript |
MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities |
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.
Python Integrations
| Integration | Package | Version Badge |
|---|---|---|
| OpenAI | openinference-instrumentation-openai |
|
| OpenAI Agents | openinference-instrumentation-openai-agents |
|
| LlamaIndex | openinference-instrumentation-llama-index |
|
| DSPy | openinference-instrumentation-dspy |
|
| AWS Bedrock | openinference-instrumentation-bedrock |
|
| LangChain | openinference-instrumentation-langchain |
|
| MistralAI | openinference-instrumentation-mistralai |
|
| Google GenAI | openinference-instrumentation-google-genai |
|
| Google ADK | openinference-instrumentation-google-adk |
|
| Guardrails | openinference-instrumentation-guardrails |
|
| VertexAI | openinference-instrumentation-vertexai |
|
| CrewAI | openinference-instrumentation-crewai |
|
| Haystack | openinference-instrumentation-haystack |
|
| LiteLLM | openinference-instrumentation-litellm |
|
| Groq | openinference-instrumentation-groq |
|
| Instructor | openinference-instrumentation-instructor |
|
| Anthropic | openinference-instrumentation-anthropic |
|
| Smolagents | openinference-instrumentation-smolagents |
|
| Agno | openinference-instrumentation-agno |
|
| MCP | openinference-instrumentation-mcp |
|
| Pydantic AI | openinference-instrumentation-pydantic-ai |
|
| Autogen AgentChat | openinference-instrumentation-autogen-agentchat |
|
| Portkey | openinference-instrumentation-portkey |
| Integration | Package | Version Badge |
|---|---|---|
| OpenAI | @arizeai/openinference-instrumentation-openai |
|
| LangChain.js | @arizeai/openinference-instrumentation-langchain |
|
| Vercel AI SDK | @arizeai/openinference-vercel |
|
| BeeAI | @arizeai/openinference-instrumentation-beeai |
|
| Mastra | @arizeai/openinference-mastra |
| Integration | Package | Version Badge |
|---|---|---|
| LangChain4j | openinference-instrumentation-langchain4j |
|
| SpringAI | openinference-instrumentation-springAI |
| Platform | Description | Docs |
|---|---|---|
| BeeAI | AI agent framework with built-in observability | Integration Guide |
| Dify | Open-source LLM app development platform | Integration Guide |
| Envoy AI Gateway | AI Gateway built on Envoy Proxy for AI workloads | Integration Guide |
| LangFlow | Visual framework for building multi-agent and RAG applications | Integration Guide |
| LiteLLM Proxy | Proxy server for LLMs | Integration Guide |
Join our community to connect with thousands of AI builders.
See the migration guide for a list of breaking changes.
Copyright 2025 Arize AI, Inc. All Rights Reserved.
Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.
This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.
Please log in to share your review and rating for this MCP.
{
"mcpServers": {
"phoenix-mcp": {
"command": "npx",
"args": [
"-y",
"@arizeai/phoenix-mcp"
],
"env": {
"API_KEY": "<YOUR_API_KEY>"
}
}
}
}claude mcp add phoenix-mcp npx -y @arizeai/phoenix-mcpExplore related MCPs that share similar capabilities and solve comparable challenges
by netdata
Delivers real‑time, per‑second infrastructure monitoring with zero‑configuration agents, on‑edge machine‑learning anomaly detection, and built‑in dashboards.
by msgbyte
Provides integrated website traffic analysis, uptime checking, and server health monitoring in a single self‑hosted platform.
by grafana
Provides programmatic access to a Grafana instance and its surrounding ecosystem through the Model Context Protocol, enabling AI assistants and other clients to query and manipulate dashboards, datasources, alerts, incidents, on‑call schedules, and more.
by dynatrace-oss
Provides a local server that enables real‑time interaction with the Dynatrace observability platform, exposing tools for querying data, retrieving problems, sending Slack notifications, and integrating AI assistance.
by pydantic
Provides tools to retrieve and query OpenTelemetry trace and metric data from Pydantic Logfire, allowing LLMs to analyze distributed traces and run arbitrary SQL queries against telemetry records.
by VictoriaMetrics-Community
Provides a Model Context Protocol server exposing read‑only VictoriaMetrics APIs, enabling seamless monitoring, observability, and automation through AI‑driven assistants.
by GeLi2001
Enables interaction with the Datadog API through a Model Context Protocol server, providing access to monitors, dashboards, metrics, logs, events, and incident data.
by last9
Provides AI agents with real‑time production context—including logs, metrics, traces, and alerts—through a Model Context Protocol server, enabling automatic code fixing and faster debugging.
by metoro-io
Provides an MCP server that exposes Metoro's eBPF‑based telemetry APIs to large language models, enabling AI‑driven queries and insights about Kubernetes clusters.