Arize Phoenix

Arize Phoenix is an open-source LLM observability platform for tracing, evaluation, and debugging.

Features

  • LLM tracing - Trace all LLM calls with full context
  • Evaluation - Built-in LLM evaluators for quality assessment
  • Embedding analysis - Visualize and debug embeddings
  • Dataset management - Create and manage evaluation datasets
  • Open source - Self-host or use Arize cloud

Quick start

Run Phoenix locally:

pip install arize-phoenix
phoenix serve

Or with Docker:

docker run -p 6006:6006 arizephoenix/phoenix:latest

Access Phoenix at http://localhost:6006.

Configuration

Phoenix accepts OpenTelemetry traces natively:

# otel-collector-config.yaml
receivers:
  otlp:
    protocols:
      grpc:
        endpoint: 0.0.0.0:4317

exporters:
  otlp:
    endpoint: http://localhost:4317
    tls:
      insecure: true

service:
  pipelines:
    traces:
      receivers: [otlp]
      exporters: [otlp]

Configure Agent Gateway:

config:
  tracing:
    otlpEndpoint: http://localhost:4317
    randomSampling: true

binds:
- port: 3000
  listeners:
  - routes:
    - backends:
      - ai:
          name: openai
          provider:
            openAI:
              model: gpt-4o-mini
      policies:
        backendAuth:
          key: "$OPENAI_API_KEY"

Docker Compose example

version: '3'
services:
  agentgateway:
    image: ghcr.io/agentgateway/agentgateway:latest
    ports:
      - "3000:3000"
    volumes:
      - ./config.yaml:/etc/agentgateway/config.yaml
    environment:
      - OTEL_EXPORTER_OTLP_ENDPOINT=http://phoenix:4317

  phoenix:
    image: arizephoenix/phoenix:latest
    ports:
      - "6006:6006"
      - "4317:4317"

Learn more