For the complete documentation index, see llms.txt. Markdown versions of all docs pages are available by appending .md to any docs URL.
LangSmith
Integrate agentgateway with LangSmith for LLM debugging and monitoring
LangSmith is LangChain’s platform for debugging, testing, evaluating, and monitoring LLM applications.
Features
- Trace logging - Detailed request/response logging
- Debugging - Step-through debugging of LLM calls
- Evaluation - Automated testing and evaluation
- Monitoring - Production monitoring and alerting
- Datasets - Build and manage evaluation datasets
Setup
- Sign up at smith.langchain.com
- Create a project and get your API key
- Configure the OpenTelemetry Collector to forward traces
Configuration
LangSmith accepts OpenTelemetry traces via their OTLP endpoint:
# otel-collector-config.yaml
receivers:
otlp:
protocols:
grpc:
endpoint: 0.0.0.0:4317
exporters:
otlphttp:
endpoint: https://api.smith.langchain.com/otel
headers:
x-api-key: "${LANGSMITH_API_KEY}"
service:
pipelines:
traces:
receivers: [otlp]
exporters: [otlphttp]Configure agentgateway:
# yaml-language-server: $schema=https://agentgateway.dev/schema/config
config:
tracing:
otlpEndpoint: http://localhost:4317
randomSampling: true
binds:
- port: 3000
listeners:
- routes:
- backends:
- ai:
name: openai
provider:
openAI:
model: gpt-4o-mini
policies:
backendAuth:
key: "$OPENAI_API_KEY"