What is Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting.?
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides tracing, evaluation, datasets, experiments, a playground, and prompt management. It is vendor and language agnostic with support for popular frameworks and LLM providers.
Documentation
Phoenix is an open-source AI observability platform designed for experimentation, evaluation, and troubleshooting. It provides:
Tracing - Trace your LLM application's runtime using OpenTelemetry-based instrumentation.
Evaluation - Leverage LLMs to benchmark your application's performance using response and retrieval evals.
Datasets - Create versioned datasets of examples for experimentation, evaluation, and fine-tuning.
Experiments - Track and evaluate changes to prompts, LLMs, and retrieval.
Phoenix runs practically anywhere, including your local machine, a Jupyter notebook, a containerized deployment, or in the cloud.
Installation
Install Phoenix via pip or conda
pip install arize-phoenix
Phoenix container images are available via Docker Hub and can be deployed using Docker or Kubernetes. Arize AI also provides cloud instances at app.phoenix.arize.com.
Packages
The arize-phoenix package includes the entire Phoenix platfom. However if you have deployed the Phoenix platform, there are light-weight Python sub-packages and TypeScript packages that can be used in conjunction with the platfrom.
MCP server implementation for Arize Phoenix providing unified interface to Phoenix's capabilities
Tracing Integrations
Phoenix is built on top of OpenTelemetry and is vendor, language, and framework agnostic. For details about tracing integrations and example applications, see the OpenInference project.