- Explore MCP Servers
- phoenix-mcp
Arize Phoenix MCP Server
What is Arize Phoenix MCP Server
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting of machine learning models, particularly focusing on large language models (LLMs). It supports tracing, evaluation, and dataset management, making it a comprehensive tool for developers and data scientists to analyze and optimize their AI applications.
Use cases
Phoenix is utilized for a variety of AI-related tasks including tracing LLM applications to monitor their performance in real-time, evaluating model outputs through response and retrieval assessments, managing datasets for fine-tuning and experimentation, and conducting structured experiments to compare model performance. Users can also optimize prompts in a controlled environment.
How to use
To install Phoenix, you can use package managers such as pip or conda. For pip, run pip install arize-phoenix
. Alternatively, Phoenix can be deployed using Docker containers or Kubernetes. Once installed, you can access various features such as tracing, evaluation, and prompt management through provided APIs and integrated frameworks.
Key features
Key features of Phoenix include OpenTelemetry-based tracing for monitoring and debugging, evaluation tools for assessing model performance, version-controlled datasets for experimentation, systematic management of prompts, and integrations with popular frameworks. It is also vendor and language agnostic, allowing broad applicability across various development environments.
Where to use
Phoenix can be employed in diverse environments such as local setups, Jupyter notebooks, or cloud-based applications. It is compatible with containerized deployments using Docker or Kubernetes, making it suitable for both individual developers and enterprise-level integrations.
Overview
What is Arize Phoenix MCP Server
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting of machine learning models, particularly focusing on large language models (LLMs). It supports tracing, evaluation, and dataset management, making it a comprehensive tool for developers and data scientists to analyze and optimize their AI applications.
Use cases
Phoenix is utilized for a variety of AI-related tasks including tracing LLM applications to monitor their performance in real-time, evaluating model outputs through response and retrieval assessments, managing datasets for fine-tuning and experimentation, and conducting structured experiments to compare model performance. Users can also optimize prompts in a controlled environment.
How to use
To install Phoenix, you can use package managers such as pip or conda. For pip, run pip install arize-phoenix
. Alternatively, Phoenix can be deployed using Docker containers or Kubernetes. Once installed, you can access various features such as tracing, evaluation, and prompt management through provided APIs and integrated frameworks.
Key features
Key features of Phoenix include OpenTelemetry-based tracing for monitoring and debugging, evaluation tools for assessing model performance, version-controlled datasets for experimentation, systematic management of prompts, and integrations with popular frameworks. It is also vendor and language agnostic, allowing broad applicability across various development environments.
Where to use
Phoenix can be employed in diverse environments such as local setups, Jupyter notebooks, or cloud-based applications. It is compatible with containerized deployments using Docker or Kubernetes, making it suitable for both individual developers and enterprise-level integrations.
Content
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
- Tracing - Trace your LLM applicationβs runtime using OpenTelemetry-based instrumentation.
- Evaluation - Leverage LLMs to benchmark your applicationβs performance using response and retrieval evals.
- Datasets - Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.
- Experiments - Track and evaluate changes to prompts, LLMs, and retrieval.
- Playground- Optimize prompts, compare models, adjust parameters, and replay traced LLM calls.
- Prompt Management- Manage and test prompt changes systematically using version control, tagging, and experimentation.
Phoenix is vendor and language agnostic with out-of-the-box support for popular frameworks (π¦LlamaIndex, π¦βLangChain, Haystack, π§©DSPy, π€smolagents) and LLM providers (OpenAI, Bedrock, MistralAI, VertexAI, LiteLLM, Google GenAI and more). For details on auto-instrumentation, check out the OpenInference project.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Installation
Install Phoenix via pip
or conda
pip install arize-phoenix
Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes.
Packages
The arize-phoenix
package includes the entire Phoenix platfom. However if you have deployed the Phoenix platform, there are light-weight Python sub-packages and TypeScript packages that can be used in conjunction with the platfrom.
Subpackages
Package | Language | Description |
---|---|---|
arize-phoenix-otel | Python |
Provides a lightweight wrapper around OpenTelemetry primitives with Phoenix-aware defaults |
arize-phoenix-client | Python |
Lightweight client for interacting with the Phoenix server via its OpenAPI REST interface |
arize-phoenix-evals | Python |
Tooling to evaluate LLM applications including RAG relevance, answer relevance, and more |
@arizeai/phoenix-client | JavaScript |
Client for the Arize Phoenix API |
@arizeai/phoenix-mcp | JavaScript |
MCP server implementation for Arize Phoenix providing unified interface to Phoenixβs capabilities |
Tracing Integrations
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.
Python Integrations
Integration | Package | Version Badge |
---|---|---|
OpenAI | openinference-instrumentation-openai |
|
OpenAI Agents | openinference-instrumentation-openai-agents |
|
LlamaIndex | openinference-instrumentation-llama-index |
|
DSPy | openinference-instrumentation-dspy |
|
AWS Bedrock | openinference-instrumentation-bedrock |
|
LangChain | openinference-instrumentation-langchain |
|
MistralAI | openinference-instrumentation-mistralai |
|
Google GenAI | openinference-instrumentation-google-genai |
|
Google ADK | openinference-instrumentation-google-adk |
|
Guardrails | openinference-instrumentation-guardrails |
|
VertexAI | openinference-instrumentation-vertexai |
|
CrewAI | openinference-instrumentation-crewai |
|
Haystack | openinference-instrumentation-haystack |
|
LiteLLM | openinference-instrumentation-litellm |
|
Groq | openinference-instrumentation-groq |
|
Instructor | openinference-instrumentation-instructor |
|
Anthropic | openinference-instrumentation-anthropic |
|
Smolagents | openinference-instrumentation-smolagents |
|
Agno | openinference-instrumentation-agno |
|
MCP | openinference-instrumentation-mcp |
|
Pydantic AI | openinference-instrumentation-pydantic-ai |
|
Autogen AgentChat | openinference-instrumentation-autogen-agentchat |
|
Portkey | openinference-instrumentation-portkey |
JavaScript Integrations
Integration | Package | Version Badge |
---|---|---|
OpenAI | @arizeai/openinference-instrumentation-openai |
|
LangChain.js | @arizeai/openinference-instrumentation-langchain |
|
Vercel AI SDK | @arizeai/openinference-vercel |
|
BeeAI | @arizeai/openinference-instrumentation-beeai |
|
Mastra | @arizeai/openinference-mastra |
Platforms
Phoenix has native integrations with LangFlow, LiteLLM Proxy, and BeeAI.
Community
Join our community to connect with thousands of AI builders.
- π Join our Slack community.
- π Read our documentation.
- π‘ Ask questions and provide feedback in the #phoenix-support channel.
- π Leave a star on our GitHub.
- π Report bugs with GitHub Issues.
- π Follow us on π.
- πΊοΈ Check out our roadmap to see where weβre heading next.
- π§βπ« Deep dive into everything Agents and LLM Evaluations on Arizeβs Learning Hubs.
Breaking Changes
See the migration guide for a list of breaking changes.
Copyright, Patent, and License
Copyright 2025 Arize AI, Inc. All Rights Reserved.
Portions of this code are patent protected by one or more U.S. Patents. See the IP_NOTICE.
This software is licensed under the terms of the Elastic License 2.0 (ELv2). See LICENSE.