Skip to main content

Documentation Index

Fetch the complete documentation index at: https://arize-ax.mintlify.dev/docs/llms.txt

Use this file to discover all available pages before exploring further.

What’s New

October 11, 2023

Prompt Playground

Uncover poorly performing prompt templates, iterate on them in real time and verify improved LLM outputs before deployment. Choose to edit the prompt variables, LLM parameters, or the prompt template itself. Then re-run the prompt to easily compare responses for analysis prior to implementing a new or revised prompt template. Learn more here.

Enhancements

October 26, 2023

VertexAI Integration for Prompt Playground

Users can now iterate on prompts in the Prompt Playground with Google’s models built on PaLM 2. This integration allows users to iterate on prompt templates, parameters, and variables in the platform and compare responses. Additionally, users can now compare LLM providers by comparing prompt runs between OpenAI and VertexAI. Learn more ->

Monitor Downtime

Users can now schedule monitor downtime - a recurring time period in which monitors should not be evaluated. Monitor downtime aims to reduce noise in monitors due to periods of unstable or low traffic. Learn more ->

Monitors Widget in Dashboards

“Alert Graph” is a new type of dashboard widget, enabling users to view historical alerting data alongside other key metrics in their dashboard. Simply select “Alert Graph” from the widgets menu.

Text Widgets In Dashboards

Text Widgets are now included in dashboards. Users can create text widgets to share notes and observations.

GraphQL Usability Enhancement for Monitors

Users can now use spaceId + modelName to create a monitor in lieu of using modelID, allowing users to streamline monitor creation. To use, simply replace the modelId field with spaceId (from url) and the modelName (the same field used for create a file import job).

Support for gpt-3.5-instruct-turbo

Users can now select to use gpt-3.5-instruct-turbo as their LLM of choice when using the Prompt Playground, or summarizing embeddings clusters.

Interactive Calibration Chart and New Precision Recall (PR) Curve

Users can interact with the calibration chart in the “More Charts” view of Performance Tracing. The chart now includes traffic bars and users have the ability to zoom into certain regions of their data, allowing for more granular insight. In addition, users can now select to view a PR curve for further analysis.

Y-Axis Controls On Dashboards

Users can now customize their Y-axis domain, customizing the min and max per-widget. Simply go into the edit form of a line or bar chart widget and hit the Display tab to control. Alternatively, users can employ auto and the charting library will compute what it believes to be the most comprehensible domain automatically.

Simultaneous UMAP visualization for multiple embedding features

Visualize your UPMAP for multiple embedding features for a multi-input model simultaneously, without needing to increase the number of predictions sent. To view multiple embedding features, select them in the “Dimensions” dropdown.

”Average” Option Added to Quantiles Plots

For additional views, users can now plot the mean, in addition to the previously offered median in quantile plots.

Python SDK v7.6.1

  • Added validation on embedding raw data for batch and record-at-a-time loggers.
  • Raised the validation string limits for string fields allowing users to send much larger strings.
  • Added truncation warnings if strings reach size limit.
Learn about Python SDK fixes and improvements here.

Python SDK v7.5.1

  • Core SDK is compatible with Python 3.6+ (previously 3.8+)
  • SDK now works with any version of pyarrow
Learn about Python SDK fixes and improvements here.

📚 New Content

LLM Tracing and Observability with Arize Phoenix How Wayfair Achieves Self-Service Onboarding for MLOps Tooling

Anyscale + Arize

We’re proud to be a launch partner of Anyscale Endpoints! Through an early integration, Arize and Anyscale Endpoints enable developers to achieve LLM observability across different use cases on any cloud as their AI applications evolve.