Install
Setup
Use the register function to connect your application to Phoenix:Run Portkey
By instrumenting Portkey, spans will be created whenever requests are made through the AI Gateway and will be sent to the Phoenix server for collection.Basic Usage with OpenAI
Using Portkey SDK Directly
Observe
Now that you have tracing setup, all requests through Portkey’s AI Gateway will be streamed to your running Phoenix instance for observability and evaluation. You’ll be able to see:- Request/Response Traces: Complete visibility into LLM interactions
- Routing Decisions: Which provider was selected and why
- Fallback Events: When and why fallbacks were triggered
- Cache Performance: Hit/miss rates and response times
- Cost Tracking: Token usage and costs across providers
- Latency Metrics: Response times for each provider and route

