Loading...
Phoenix v10.5.0 now supports Deepseek and xAI models in Playground natively. Previous versions of Phoenix supported these as custom model endpoints, but that process has now been streamlined to offer these model providers from the main Playground dropdown.
We've added a host of new methods to the JS client:
getExperiment - allows you to retrieve an Experiment to view its results, and run evaluations on it
evaluateExperiment - allows you to evaluate previously run Experiments using LLM as a Judge or Code-based evaluators
createDataset - allows you to create Datasets in Phoenix using the client
appendDatasetExamples - allows you to append additional examples to a Dataset
You can now run Experiments using the Phoenix JS client! Use Experiments to test different iterations of your applications over a set of test cases, then evaluate the results.
This release includes:
Native tracing of tasks and evaluators
Async concurrency queues
Support for any evaluator (including bring your own evals)
Phoenix v9.0.0 release brings major updates to annotation support, and a whole host of other improvements.
Up until now, Phoenix has only supported one annotation of a given type on each trace. We've now unlocked that limit, allowing you to capture multiple values of an annotation label on each span.
In addition, we've added:
API support for annotations - create, query, and update annotations through the REST API
Additional support for code evaluations as annotations
Support for arbitrary metadata on annotations
Annotation configurations to structure your annotations within and across projects
Now you can create custom global and per-project data retention polices to remove traces after a certain window of time, or based on number of traces. Additionally, you can now view your disk usage in the Settings page of Phoenix.
We've added hotkeys to Phoenix!
You can now use j
and k
to quickly page through your traces, and e
and n
to add annotations and notes - you never have to lift your hands off the keyboard again!
We’ve added a Python auto-instrumentation library for the Google GenAI SDK. This enables seamless tracing of GenAI workflows with full OpenTelemetry compatibility. Traces can be exported to any OpenTelemetry collector.
For more details on how to set up the tracing integration seamlessly:
Additionally, the Google GenAI instrumentor is now supported and works seamlessly with Span Replay in Phoenix, enabling deep trace inspection and replay for more effective debugging and observability.
Big thanks to Harrison Chu for his contributions.
Loading...
Available in Phoenix 8.29+
Phoenix now supports Transport Layer Security (TLS) for both HTTP and gRPC connections, enabling encrypted communication and optional mutual TLS (mTLS) authentication. This enhancement provides a more secure foundation for production deployments.
Secure HTTP & gRPC Connections: Phoenix can now serve over HTTPS and secure gRPC.
Flexible TLS Configuration: TLS settings are managed via environment variables.
Optional Client Verification: Support for mTLS with configurable client certificate validation.
Improved Testing: TLS-aware infrastructure added to integration tests.
Better Visibility: Server startup logs now display TLS status.
Set the following environment variables to enable and customize TLS:
Available in Phoenix 8.28+
When stopping the Phoenix server via Ctrl+C
, the shutdown process now exits cleanly without displaying a traceback or returning a non-zero exit code. Previously, a KeyboardInterrupt
and CancelledError
traceback could appear, ending the process with status code 130. The server now swallows the interrupt for a smoother shutdown experience, exiting with code 0 by default to reflect intentional termination.
Available in Phoenix 8.27+
Improved trace navigation by automatically scrolling the selected span into view when a user navigates to a specific trace. This enhancement eliminates the need for manual searching or scrolling, allowing users to immediately focus on the span of interest. It's especially useful when navigating from links or alerts that point to a specific span, improving debugging efficiency. This change contributes to a smoother and more intuitive trace exploration experience.
Loading...
Loading...
Available in Phoenix 8.25+
Loading...
Loading...
Available in Phoenix 8.22+
We’ve added support for Prompt Tagging in the Phoenix client. This new feature gives you more control and visibility over your prompts throughout the development lifecycle.
Tag prompts directly in your code and see those tags reflected in the Phoenix UI.
Label prompt versions as development
, staging
, or production
— or define your own custom tags.
Add tag descriptions to provide additional context or list out all tags.
Loading...
Available in Phoenix 8.20+
Newly added to the OpenAI Agent SDK is support for MCP Span Info, allowing for the tracing and extraction of useful information about MCP tool listings. Use the Phoenix OpenAI Agents SDK for powerful agent tracing.
Loading...
Loading...
Loading...
Available in Phoenix 8.19+
Available in Phoenix 8.15+
Available in Phoenix 8.14+
Loading...
Loading...
Loading...
Loading...
Loading...
Available in Phoenix 8.0+
Phoenix has made it even simpler to get started with tracing by introducing one-line auto-instrumentation. By using register(auto_instrument=True)
, you can enable automatic instrumentation in your application, which will set up instrumentors based on your installed packages.
Loading...
Loading...
Loading...
Loading...
Available in Phoenix 4.11+
Our integration with Guardrails AI allows you to capture traces on guard usage and create datasets based on these traces. This integration is designed to enhance the safety and reliability of your LLM applications, ensuring they adhere to predefined rules and guidelines.
Loading...
Loading...
Loading...
: Use Float for token count summaries
: Improve browser compatibility for table sizing
: Simplify homeLoaderQuery
to prevent idle timeout errors
: Add /readyz endpoint to confirm database connectivity
: Allow scroll on settings page
Tool call and result IDs are now shown in the span details view. Each ID is placed within a collapsible header and can be easily copied. This update also supports spans with multiple tool calls. Get started with tracing your tool calls .
Check out documentation on .
: Add aiohttp to container for azure-identity
You can now delete experiments directly from the action menu, making it quicker to manage and clean up your workspace. This update streamlines experiment management by reducing the steps needed to remove outdated or unnecessary runs. Get started with experiments .
: Show the date format in the explanation
In the New Project tab, we've added quick setup to instrument your application for , , and the .
Easily configure all integrations with streamlined instructions. Check out all Phoenix here.
: Remove shadow on button group
: Fixed broken popovers
For more details, you can check the docs and explore further options.
Check out the here.
PHOENIX_TLS_ENABLED
boolean
Enable or disable TLS (true
/false
)
PHOENIX_TLS_CERT_FILE
string
Path to TLS certificate file
PHOENIX_TLS_KEY_FILE
string
Path to private key file
PHOENIX_TLS_KEY_FILE_PASSWORD
string
Password for encrypted private key file
PHOENIX_TLS_CA_FILE
string
Path to CA certificate (for client verification)
PHOENIX_TLS_VERIFY_CLIENT
boolean
Enable client cert verification
Available in Phoenix 8.26+
Available in Phoenix 8.5+
We’ve introduced several enhancements to Projects, providing greater flexibility and control over how you interact with data. These updates include:
Custom Time Ranges: You can now specify custom time ranges to filter traces and spans.
Root Span Filter for Spans: Improved filtering options allow you to filter by root spans, helping to isolate and debug issues more effectively.
Available in Phoenix 8.11+
Save and Load from Prompts: You can now save and load configurations directly from prompts.
Save and Load from Default Model Config: Default model configurations can be saved and loaded.
Budget Token Management: Added the ability to adjust the budget token value.
Thinking Configuration Toggle: You can now enable or disable the “thinking” feature.
Important Note: The default model config does not automatically apply to saved prompts. To include default thinking settings, ensure they are saved within the specific prompt.
Available in Phoenix 8.26+
This release introduces automatic distributed tracing for Anthropic’s Model Context Protocol (MCP). Using OpenTelemetry, you can now:
Propagate context across MCP client-server boundaries
Generate end-to-end traces of your AI system across services and languages
Gain full visibility into how models access and use external context
The openinference-instrumentation-mcp
package handles this for you by:
Creating spans for MCP client operations
Injecting trace context into MCP requests
Extracting and continuing the trace context on the server
Associating the context with OTEL spans on the server side
Instrument both MCP client and server with OpenTelemetry.
Add the openinference-instrumentation-mcp
package.
Spans will propagate across services, appearing as a single connected trace in Phoenix.
Full example usage is available:
Big thanks to Adrian Cole and Anuraag Agrawal for their contributions to this feature.
Available in Phoenix 8.24+
This update enhances the Project Management API with more flexible project identification:
Enhanced project identification: Added support for identifying projects by both ID and hex-encoded name and introduced a new _get_project_by_identifier
helper function
Also includes streamlined operations, better validation & error handling, and expanded test coverage.
Available in Phoenix 8.0+
Phoenix prompt management will now let you create, modify, tag, and version control prompts for your applications. Some key highlights from this release:
Versioning & Iteration: Seamlessly manage prompt versions in both Phoenix and your codebase.
New TypeScript Client: Sync prompts with your JavaScript runtime, now with native support for OpenAI, Anthropic, and the Vercel AI SDK.
New Python Client: Sync templates and apply them to AI SDKs like OpenAI, Anthropic, and more.
Standardized Prompt Handling: Native normalization for OpenAI, Anthropic, Azure OpenAI, and Google AI Studio.
Enhanced Metadata Propagation: Track prompt metadata on Playground spans and experiment metadata in dataset runs.
Available in Phoenix 8.19+
Within each project, there is now a Config tab to enhance customization. The default tab can now be set per project, ensuring the preferred view is displayed.
Experiments: Included delete experiment option to action menu
Feature: Added support for specifying admin users via an environment variable at startup
Annotation: Now displays metadata
Settings Page: Now split across tabs for improved navigation and easier access
Feedback: Added full metadata
Projects: Improved performance
UI: Added date format descriptions to explanations
Available in Phoenix 4.6+
Datasets: Datasets are a new core feature in Phoenix that live alongside your projects. They can be imported, exported, created, curated, manipulated, and viewed within the platform, and should make a few flows much easier:
Fine-tuning: You can now create a dataset based on conditions in the UI, or by manually choosing examples, then export these into csv or jsonl formats readymade for fine-tuning APIs.
Experimentation: External datasets can be uploaded into Phoenix to serve as the test cases for experiments run in the platform.
Experiments: Our new Datasets and Experiments feature enables you to create and manage datasets for rigorous testing and evaluation of your models. You can now run comprehensive experiments to measure and analyze the performance of your LLMs in various scenarios.
Available in Phoenix 8.17+
You can now specify one or more admin users at startup using an environment variable. This is especially useful for managed deployments, allowing you to define admin access in a manifest or configuration file. The specified users will be automatically seeded into the database, enabling immediate login without manual setup.
Phoenix is now available for deployment as a fully hosted service.
With hosted instances, your data is stored between sessions, and you can easily share your work with team members.
We are partnering with LlamaIndex to power a new observability platform in LlamaCloud: LlamaTrace. LlamaTrace will automatically capture traces emitted from your LlamaIndex applications, and store them in a persistent, cloud- accessible Phoenix instance.
Available in Phoenix 8.21+
Available in Phoenix 8.13+
We've introduced the OpenAI Agents SDK for Python which provides enhanced visibility into agent behavior and performance.
Installation
Includes an OpenTelemetry Instrumentor that traces agents, LLM calls, tool usage, and handoffs.
With minimal setup, use the register
function to connect your app to Phoenix and view real-time traces of agent workflows.
For more details on a quick setup, check out our integration documentation:
Available in Phoenix 8.23+
This release introduces a REST API for managing projects, complete with full CRUD functionality and access control. Key features include:
CRUD Operations: Create, read, update, and delete projects via the new API endpoints.
Role-Based Access Control:
Admins can create, read, update, and delete projects
Members can create and read projects, but cannot modify or delete them.
Additional Safeguards: Immutable Project Names, Default Project Protection, Comprehensive Integration Tests
Available in Phoenix 6.0+
Sessions allow you to group multiple responses into a single thread. Each response is still captured as a single trace, but each trace is linked together and presented in a combined view.
Added support for FastAPI and GraphQL extensions
Fixed a bug where Anthropic LLM as a Judge responses would be labeled as unparseable
Fixed a bug causing 500 errors on client.get_traces_dataset() and client.get_spans_dataframe()
Added the ability for authentication to work from behind a proxy
Added an environment variable to set default admin passwords in auth
Available in Phoenix 5.0+
We've added Authentication and Rules-based Access Controls to Phoenix. This was a long-requested feature set, and we're excited for the new uses of Phoenix this will unlock!
The auth feature set includes:
Secure Access: All of Phoenix’s UI & APIs (REST, GraphQL, gRPC) now require access tokens or API keys. Keep your data safe!
RBAC (Role-Based Access Control): Admins can manage users; members can update their profiles—simple & secure.
API Keys: Now available for seamless, secure data ingestion & querying.
OAuth2 Support: Easily integrate with Google, AWS Cognito, or Auth0. ✉ Password Resets via SMTP to make security a breeze.
Numerous stability improvements to our hosted Phoenix instances accessed on app.phoenix.arize.com
Added a new command to easily launch a Phoenix client from the cli: phoenix serve
Implemented simple email sender to simplify dependencies
Improved error handling for imported spans
Replaced hdbscan with fast-hdbscan Added PHOENIX_CSRF_TRUSTED_ORIGINS environment variable to set trusted origins
Added support for Mistral 1.0
Fixed an issue that caused px.Client().get_spans_dataframe() requests to time out
Available in Phoenix 7.9+
In addition to using our automatic instrumentors and tracing directly using OTEL, we've now added our own layer to let you have the granularity of manual instrumentation without as much boilerplate code.
You can now access a tracer object with streamlined options to trace functions and code blocks. The main two options are:
Using the decorator @tracer.chain
traces the entire function automatically as a Span in Phoenix. The input, output, and status attributes are set based on the function's parameters and return value.
Using the tracer in a with
clause allows you to trace specific code blocks within a function. You manually define the Span name, input, output, and status.
Available in Phoenix 7.0+
Sessions allow you to group multiple responses into a single thread. Each response is still captured as a single trace, but each trace is linked together and presented in a combined view.
Prompt Playground: Added support for arbitrary string model names Added support for Gemini 2.0 Flash Improved template editor ergonomics
Evals: Added multimodal message template support
Tracing: Added JSON pretty printing for structured data outputs (thank you sraibagiwith100x!) Added a breakdown of token types in project summary
Bug Fixes: Changed trace latency to be computed every time, rather than relying on root span latency, Added additional type checking to handle non-string values when manually instrumenting (thank you Manuel del Verme!)
Available in Phoenix 8.30+
The Phoenix client now includes the SpanQuery
DSL, enabling more advanced and flexible span querying for distributed tracing and telemetry data. This allows users to perform complex queries on span data, improving trace analysis and debugging.
In addition, the get_spans_dataframe
method has been migrated, offering an easy-to-use way to extract span-related information as a Pandas DataFrame. This simplifies data processing and visualization, making it easier to analyze trace data within Python-based environments.
: Add load more and loading state to the infinite scroll
: Hide menu for changing role for self in UsersTable
: Prevent admins from changing their own roles
: Remove WebSocket dependency and migrate to Multipart Subscriptions
: Your selected columns will now remain consistent across sessions, ensuring a more seamless workflow.
: Easily filter data directly from the table view using metadata attributes.
: Quickly apply common metadata filters for faster navigation.
: Major speed improvements in project tracing views & visibility into database usage in settings
: Query to get number of spans for each trace
: Show + n
more spans in trace table
: Add Token component
: Remove double fetching of spans
: Don't fetch new traces when the traces slideover is visible
: Fix scrolling on trace tree
: Added annotations to experiment JSON downloads
: Add none
as option for tool choice for anthropic 0.49.0
: Port slider component to react-aria
We’re excited to announce a powerful capability in the OSS library openinference-instrumentation-mcp
— seamless OTEL context propagation for MCP clients and servers.
: Restore streaming
: update Gemini models
: Route user to forgot-password page in welcome email url
Check out the docs and this for more on prompts.📝
Learn more in .
: Use correlated subquery for orphan spans
: Add toggle to treat orphan spans as root
: Upgrade react-router, vite, vitest
For more details on using datasets see our or .
For more details, check out our full .
: Smaller page sizes
: Improved performance on projects page
: Allow hover anywhere on experiment cell
: Show metadata
: Show full metadata
In addition to our existing notebook, CLI, and self-hosted deployment options, we’re excited to announce that Phoenix is now available as a .
Hosted Phoenix is 100% free-to-use, .
: Allow the option to have no configured working directory when using Postgres
: Cache project table results when toggling the details slide-over for improved performance
: Add chat and message components for note-taking
You can now toggle the option to treat orphan spans as root when viewing your spans. Additionally, we've enhanced the UI with an icon view in span details for better visibility in smaller displays. Learn more in our .
: Disable streaming when a dialog is open
: Removed unpredictable playground transformations
: Azure API key made optional, included specialized UI for thinking budget parameter
: Make the spans table the default tab
: Added react-aria Tabs components
: Download experiment runs and annotations as CSV
Check out our to test these features.
: add PHOENIX_ALLOWED_ORIGINS env
: Delete annotations in the feedback table, Make feedback table scrollable
: Allow scrolling the entire experiment compare table
: Make time range selector more accessible
: Don't close model settings dialog when picking Azure version
: improve PostgreSQL error message in launch_app
Sessions make it easier to visual multi-turn exchanges with your chatbot or agent Sessions launches with Python and TS/JS support. For more on sessions, check out and the .
For all the details on authentication, view our .
Prompt Playground: Now supports & Anthropic Sonnet 3.7 and Thinking Budgets
Instrumentation: to trace smolagents by Hugging Face
Integrations: Phoenix now supports &
: Show percent used of DB
: Add environment variable for allocated DB storage capacity
: Delete selected traces
: Make trace tree more readable on smaller sizes
: Ensure type is correct on run_experiment
: Allow experiment run JSON downloads
: Add anthropic thinking config param
: Add ToggleButton
Check out the for more on how to use tracer objects.
Sessions make it easier to visual multi-turn exchanges with your chatbot or agent Sessions launches with Python and TS/JS support. For more on sessions, check out and the .
: Add "Copy Name" button to project menu
: Add independent flags for whether TLS is enabled for HTTP and gRPC servers
: Log playground subscription errors
: New RBAC primitives have been introduced for FastAPI and REST APIs
Available in Phoenix 4.6+
We are introducing a new built-in function call evaluator that scores the function/tool-calling capabilities of your LLMs. This off-the-shelf evaluator will help you ensure that your models are not just generating text but also effectively interacting with tools and functions as intended.
This evaluator checks for issues arising from function routing, parameter extraction, and function generation.
Check out a .
The latest from the Phoenix team.
Available in Phoenix v10.5.0+
Deepseek and xAI models are now available in Prompt Playground!
We've added a host of new methods to the JS client:
You can now run Experiments using the Phoenix JS client! Use Experiments to test different iterations of your applications over a set of test cases, then evaluate the results.
This release includes:
Native tracing of tasks and evaluators
Async concurrency queues
Support for any evaluator (including bring your own evals)
Major Release: Phoenix v9.0.0
Available in Phoenix 9.0.0+
Phoenix's v9.0.0 release brings with it:
Customizable data retention policies
We’ve added a Python auto-instrumentation library for the Google GenAI SDK. This enables seamless tracing of GenAI workflows with full OpenTelemetry compatibility. Additionally, the Google GenAI instrumentor is now supported and works seamlessly with Span Replay in Phoenix.
Available in Phoenix 8.30+
The Phoenix client now includes the SpanQuery
DSL for more advanced span querying. Additionally, a get_spans_dataframe
method has been added to facilitate easier data extraction for span-related information.
Available in Phoenix 8.29+
Phoenix now supports Transport Layer Security (TLS) for both HTTP and gRPC connections, enabling encrypted communication and optional mutual TLS (mTLS) authentication. This enhancement provides a more secure foundation for production deployments.
Available in Phoenix 8.28+
When stopping the Phoenix server via Ctrl+C
, the shutdown process now exits cleanly with code 0 to reflect intentional termination. Previously, this would trigger a traceback with KeyboardInterrupt
, misleadingly indicating a failure.
Available in Phoenix 8.27+
Improved trace navigation by automatically scrolling the selected span into view when a user navigates to a specific trace. This enhances usability by making it easier to locate and focus on the relevant span without manual scrolling.
Available in Phoenix 8.26+
We’ve released openinference-instrumentation-mcp
, a new package in the OpenInference OSS library that enables seamless OpenTelemetry context propagation across MCP clients and servers. It automatically creates spans, injects and extracts context, and connects the full trace across services to give you complete visibility into your MCP-based AI systems.
Big thanks to Adrian Cole and Anuraag Agrawal for their contributions to this feature.
Available in Phoenix 8.26+
Phoenix now supports programmatic API key creation through a new endpoint, making it easier to automate project setup and trace logging. To enable this, set the PHOENIX_ADMIN_SECRET
environment variable in your deployment.
Available in Phoenix 8.25+
Available in Phoenix 8.24+
This update enhances the Project Management API with more flexible project identification We've added support for identifying projects by both ID and hex-encoded name and introduced a new _get_project_by_identifier
helper function.
Available in Phoenix 8.23+
Available in Phoenix 8.22+
Available in Phoenix 8.21+
The new span aside moves the Span Annotation editor into a dedicated panel, providing a clearer view for adding annotations and enhancing customization of your setup. Read this documentation to learn how annotations can be used.
Available in Phoenix 8.20+
Newly added to the OpenAI Agent SDK is support for MCP Span Info, allowing for the tracing and extraction of useful information about MCP tool listings. Use the Phoenix OpenAI Agents SDK for powerful agent tracing.
Available in Phoenix 8.20+
Available in Phoenix 8.19+
Available in Phoenix 8.17+
You can now preconfigure admin users at startup using an environment variable, making it easier to manage access during deployment. Admins defined this way are automatically seeded into the database and ready to log in.
Available in Phoenix 8.16+
You can now delete experiments directly from the action menu, making it quicker to manage and clean up your workspace.
Available in Phoenix 8.15+
Available in Phoenix 8.14+
We've added the ability to resize Span, Trace, and Session tables. Resizing preferences are now persisted in the tracing store, ensuring settings are maintained per-project and per-table.
Available in Phoenix 8.13+
Available in Phoenix 8.11+
You can now save and load configurations directly from prompts or default model settings. Additionally, you can adjust the budget token value and enable/disable the "thinking" feature, giving you more control over model behavior and resource allocation.
Available in Phoenix 8.9+
Prompt Playground now supports new GPT and Anthropic models new models with enhanced configuration options. Instrumentation options have been improved for better traceability, and evaluation capabilities have expanded to cover Audio & Multi-Modal Evaluations. Phoenix also introduces new integration support for LiteLLM Proxy & Cleanlabs evals.
Available in Phoenix 8.8+
We’ve rolled out several enhancements to Projects, offering more flexibility and control over your data. Key updates include persistent column selection, advanced filtering options for metadata and spans, custom time ranges, and improved performance for tracing views. These changes streamline workflows, making data navigation and debugging more efficient.
Available in Phoenix 8.0+
Phoenix prompt management will now let you create, modify, tag, and version control prompts for your applications. Some key highlights from this release:
Versioning & Iteration: Seamlessly manage prompt versions in both Phoenix and your codebase.
New TypeScript Client: Sync prompts with your JavaScript runtime, now with native support for OpenAI, Anthropic, and the Vercel AI SDK.
New Python Client: Sync templates and apply them to AI SDKs like OpenAI, Anthropic, and more.
Standardized Prompt Handling: Native normalization for OpenAI, Anthropic, Azure OpenAI, and Google AI Studio.
Enhanced Metadata Propagation: Track prompt metadata on Playground spans and experiment metadata in dataset runs.
Available in Phoenix 8.0+
Phoenix has made it even simpler to get started with tracing by introducing one-line auto-instrumentation. By using register(auto_instrument=True)
, you can enable automatic instrumentation in your application, which will set up instrumentors based on your installed packages.
Available in Phoenix 7.9+
In addition to using our automatic instrumentors and tracing directly using OTEL, we've now added our own layer to let you have the granularity of manual instrumentation without as much boilerplate code.
You can now access a tracer object with streamlined options to trace functions and code blocks. The main two options are using the decorator @tracer.chain
and using the tracer in a with
clause.
Available in Phoenix 7.0+
Sessions allow you to group multiple responses into a single thread. Each response is still captured as a single trace, but each trace is linked together and presented in a combined view.
Available in Phoenix 6.0+
Prompt Playground is now available in the Phoenix platform! This new release allows you to test the effects of different prompts, tools, and structured output formats to see which performs best.
Replay individual spans with modified prompts, or run full Datasets through your variations.
Easily test different models, prompts, tools, and output formats side-by-side, directly in the platform.
Available in Phoenix 5.0+
We've added Authentication and Rules-based Access Controls to Phoenix. This was a long-requested feature set, and we're excited for the new uses of Phoenix this will unlock!
Available in Phoenix 4.11.0+
Our integration with Guardrails AI allows you to capture traces on guard usage and create datasets based on these traces. This integration is designed to enhance the safety and reliability of your LLM applications, ensuring they adhere to predefined rules and guidelines.
Phoenix is now available for deployment as a fully hosted service.
We are partnering with LlamaIndex to power a new observability platform in LlamaCloud: LlamaTrace. LlamaTrace will automatically capture traces emitted from your LlamaIndex application.
Available in Phoenix 4.6+
Datasets: Datasets are a new core feature in Phoenix that live alongside your projects. They can be imported, exported, created, curated, manipulated, and viewed within the platform, and make fine-turning and experimentation easier.n
Available in Phoenix 4.6+
- allows you to retrieve an Experiment to view its results, and run evaluations on it
- allows you to evaluate previously run Experiments using LLM as a Judge or Code-based evaluators
- allows you to create Datasets in Phoenix using the client
- allows you to append additional examples to a Dataset
A host of improvements to , including one-to-many support, API access, annotation configs, and custom metadata
Hotkeys!
Tool call and result IDs are now shown in the span details view. Each ID is placed within a collapsible header and can be easily copied. This update also supports spans with multiple tool calls. Get started with tracing your tool calls .
This release introduces a REST API for managing projects, complete with full CRUD functionality and access control. Key features include CRUD Operations and Role-Based Access Control. Check out our to test these features.
We’ve added support for Prompt Tagging in the Phoenix client. This new feature gives you more control and visibility over your prompts throughout the development lifecycle. Tag prompts directly in code, label prompt versions, and add tag descriptions. Check out documentation on .
You can now toggle the option to treat orphan spans as root when viewing your spans. Additionally, we've enhanced the UI with an icon view in span details for better visibility in smaller displays. Learn more .
Within each project, there is now a Config tab to enhance customization. The default tab can now be set per project, ensuring the preferred view is displayed. Learn more in .
In the New Project tab, we've added quick setup to instrument your application for BeeAI, SmolAgents, and the OpenAI Agents SDK. Easily configure these integrations with streamlined instructions. Check out all Phoenix here.
We've introduced the OpenAI Agents SDK for Python which provides enhanced visibility into agent behavior and performance. For more details on a quick setup, check out our .
Check out docs for more.
Check out the docs and this for more on prompts!📝
Check out the for more on how to use tracer objects.
Sessions make it easier to visual multi-turn exchanges with your chatbot or agent Sessions launches with Python and TS/JS support. For more on sessions, check out and the .
Automatically capture traces as Experiment runs for later debugging. See for more information on Prompt Playground, or jump into the platform to try it out for yourself.
The auth feature set includes secure access, RBAC, API keys, and OAuth2 Support. For all the details on authentication, view our .
Check out the here.
In addition to our existing notebook, CLI, and self-hosted deployment options, we’re excited to announce that Phoenix is now available as a . With hosted instances, your data is stored between sessions, and you can easily share your work with team members.
Hosted Phoenix is 100% free-to-use, !
For more details on using datasets see our or .
Experiments: Our new Datasets and Experiments feature enables you to create and manage datasets for rigorous testing and evaluation of your models. Check out our full .
We are introducing a new built-in function call evaluator that scores the function/tool-calling capabilities of your LLMs. This off-the-shelf evaluator will help you ensure that your models are not just generating text but also effectively interacting with tools and functions as intended. Check out a .