by Arize-ai
Open-source AI observability platform enabling tracing, evaluation, dataset versioning, experiment tracking, prompt management, and interactive playground for LLM applications.
Provides end‑to‑end observability for large language model (LLM) workflows: captures runtime traces via OpenTelemetry, runs automated evaluations, stores versioned datasets, tracks experiments, and offers a UI playground for prompt engineering.
pip install arize-phoenix
Or pull the container: docker pull arizephoenix/phoenix
.arize-phoenix serve
).openinference-instrumentation-openai
, openinference-instrumentation-langchain
, etc.) or use the provided arize-phoenix-otel
package for a simplified setup.http://localhost:8080
(default) to explore traces, evaluation results, datasets, and experiments.arize-phoenix-client
(Python) or @arizeai/phoenix-client
(JS/TS) to create datasets, upload evaluations, and query metadata.arize-phoenix-evals
.Q: Do I need a cloud account to run Phoenix? A: No. Phoenix can run locally via Docker or the Python package. Cloud‑hosted instances are optional.
Q: Which frameworks are supported out of the box? A: LangChain, LlamaIndex, Haystack, DSPy, Smolagents, OpenAI SDK, Bedrock, Vertex AI, MistralAI, LiteLLM, and many more via OpenInference.
Q: How are evaluations stored? A: Evaluations are persisted as part of the Phoenix backend and can be queried through the UI or client SDKs.
Q: Can I extend the evaluation suite?
A: Yes. Custom eval functions can be added using arize-phoenix-evals
or the TypeScript counterpart.
Q: Is there a licensing cost? A: Phoenix is released under the Elastic License 2.0 and is free to use; commercial support is offered by Arize AI.
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (🦙LlamaIndex, 🦜⛓LangChain, Haystack, 🧩DSPy, 🤗smolagents) and LLM providers (OpenAI, Bedrock, MistralAI, VertexAI, LiteLLM, Google GenAI and more). For details on auto-instrumentation, check out the OpenInference project.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Install Phoenix via pip
or conda
pip install arize-phoenix
Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.
The arize-phoenix
package includes the entire Phoenix platfom. However if you have deployed the Phoenix platform, there are light-weight Python sub-packages and TypeScript packages that can be used in conjunction with the platfrom.
Package | Language | Description |
---|---|---|
arize-phoenix-otel | Python |
Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |
arize-phoenix-client | Python |
Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface |
arize-phoenix-evals | Python |
Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more |
@arizeai/phoenix-client | JavaScript |
Client for the Arize Phoenix API |
@arizeai/phoenix-evals | TypeScript |
TypeScript evaluation library for LLM applications (alpha release) |
@arizeai/phoenix-mcp | JavaScript |
MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities |
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.
Python Integrations
Integration | Package | Version Badge |
---|---|---|
OpenAI | openinference-instrumentation-openai |
|
OpenAI Agents | openinference-instrumentation-openai-agents |
|
LlamaIndex | openinference-instrumentation-llama-index |
|
DSPy | openinference-instrumentation-dspy |
|
AWS Bedrock | openinference-instrumentation-bedrock |
|
LangChain | openinference-instrumentation-langchain |
|
MistralAI | openinference-instrumentation-mistralai |
|
Google GenAI | openinference-instrumentation-google-genai |
|
Google ADK | openinference-instrumentation-google-adk |
|
Guardrails | openinference-instrumentation-guardrails |
|
VertexAI | openinference-instrumentation-vertexai |
|
CrewAI | openinference-instrumentation-crewai |
|
Haystack | openinference-instrumentation-haystack |
|
LiteLLM | openinference-instrumentation-litellm |
|
Groq | openinference-instrumentation-groq |
|
Instructor | openinference-instrumentation-instructor |
|
Anthropic | openinference-instrumentation-anthropic |
|
Smolagents | openinference-instrumentation-smolagents |
|
Agno | openinference-instrumentation-agno |
|
MCP | openinference-instrumentation-mcp |
|
Pydantic AI | openinference-instrumentation-pydantic-ai |
|
Autogen AgentChat | openinference-instrumentation-autogen-agentchat |
|
Portkey | openinference-instrumentation-portkey |
Integration | Package | Version Badge |
---|---|---|
OpenAI | @arizeai/openinference-instrumentation-openai |
|
LangChain.js | @arizeai/openinference-instrumentation-langchain |
|
Vercel AI SDK | @arizeai/openinference-vercel |
|
BeeAI | @arizeai/openinference-instrumentation-beeai |
|
Mastra | @arizeai/openinference-mastra |
Integration | Package | Version Badge |
---|---|---|
LangChain4j | openinference-instrumentation-langchain4j |
|
SpringAI | openinference-instrumentation-springAI |
Platform | Description | Docs |
---|---|---|
BeeAI | AI agent framework with built-in observability | Integration Guide |
Dify | Open-source LLM app development platform | Integration Guide |
Envoy AI Gateway | AI Gateway built on Envoy Proxy for AI workloads | Integration Guide |
LangFlow | Visual framework for building multi-agent and RAG applications | Integration Guide |
LiteLLM Proxy | Proxy server for LLMs | Integration Guide |
Join our community to connect with thousands of AI builders.
See the migration guide for a list of breaking changes.
Copyright 2025 Arize AI, Inc. All Rights Reserved.
Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.
This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.
Please log in to share your review and rating for this MCP.
{ "mcpServers": { "phoenix-mcp": { "command": "npx", "args": [ "-y", "@arizeai/phoenix-mcp" ], "env": { "API_KEY": "<YOUR_API_KEY>" } } } }
Explore related MCPs that share similar capabilities and solve comparable challenges
by grafana
Provides programmatic access to a Grafana instance and its surrounding ecosystem through the Model Context Protocol, enabling AI assistants and other clients to query and manipulate dashboards, datasources, alerts, incidents, on‑call schedules, and more.
by dynatrace-oss
Provides a local server that enables real‑time interaction with the Dynatrace observability platform, exposing tools for querying data, retrieving problems, sending Slack notifications, and integrating AI assistance.
by VictoriaMetrics-Community
Provides a Model Context Protocol server exposing read‑only VictoriaMetrics APIs, enabling seamless monitoring, observability, and automation through AI‑driven assistants.
by GeLi2001
Enables interaction with the Datadog API through a Model Context Protocol server, providing access to monitors, dashboards, metrics, logs, events, and incident data.
by QAInsights
Execute JMeter test plans through Model Context Protocol clients, capture console output, generate HTML dashboards, and automatically analyze JTL results to surface performance metrics, bottlenecks, and actionable recommendations.
by grafana
Provides a Model Context Protocol (MCP) server that enables AI agents to query Grafana Loki log data via stdin/stdout or Server‑Sent Events, supporting both local binary execution and containerized deployment.
by TocharianOU
Provides a Model Context Protocol (MCP) server that enables MCP‑compatible clients to access, search, and manage Kibana APIs using natural language or programmatic requests.
by MindscapeHQ
Provides comprehensive access to Raygun's API V3 endpoints for crash reporting and real user monitoring via the Model Context Protocol.
by grafana
Provides Model Context Protocol endpoints that enable AI assistants to query and analyze distributed tracing data stored in Grafana Tempo, supporting both stdin/stdout communication and an HTTP SSE interface.