04.2025
Our April 2025 releases
Major design refresh in Arize AX
We've refreshed Arize AX with polished fonts, spacing, color, and iconography throughout the whole platform.
Custom code evaluators
You can now run your own custom python code evaluators in Arize against your data in a secure environment. Use background tasks to run any custom code, such as URL validations, or keyword match. Learn more

Security audit logs for enterprise customers
Improve your compliance and policy adherence. You can now use audit logs to monitor data access in Arize. Note: This feature is completely opt-in and this tracking is not enabled unless a customer explicitly asks for it. Learn more
Larger dataset runs in prompt playground
We've increased the row limit for datasets in the playground, so you can run prompts in parallel on up to 100 examples.

Evaluations on experiments
You can now create and run evals on your experiments from the UI. Compare performance across different prompt templates, models, or configurations without code. Learn more →

Evaluations on experiments
You can now create and run evals on your experiments from the UI. Compare performance across different prompt templates, models, or configurations without code. Learn more →

Cancel running background tasks
When running evaluations using background tasks, you can now cancel them mid-flight while observing task logs. Learn more →

Improved UI for functions in prompt playground
We've made it easier to view, test, and validate your tool calls in prompt playground. Learn more →

Compare prompts side by side
Compare the outputs of a new prompt and the original prompt side-by-side. Tweak model parameters and compare results across your datasets.

Image segmentation support for CV models
We now support logging image segmentation to Arize. Log your segmentation coordinates and compare your predictions vs. your actuals.

New time selector on your traces
We’ve made it way easier to drill into specific time ranges, with quick presets like "last 15 minutes" and custom shorthand for specific dates and times, such as 10d
,4/1 - 4/6
, 4/1 3:00am
.
Learn more →

Prompt hub python SDK
Access and manage your prompts in code with support for OpenAI and VertexAI. Learn more
View task run history and errors
Get full visbility into your evaluation task runs, including when it ran, what triggered it, and if there were errors. Learn more →

Run evals and tasks over a date range
Easily run your online evaluation tasks over historical data.

Last updated
Was this helpful?