Quickstart: Write First Eval
This guide is a starting point for getting started with evaluations. Learn how to write and run evals in both code and the Arize UI through guided examples.
For end-to-end walkthroughs, check out our cookbooks.
1. Upload a CSV as a dataset
Download this sample CSV and upload it as a dataset in the Prompt Playground.
2. Set up a task in the playground
Load the dataset you created into prompt playground and enter the following prompt:
Who invented {attributes.output.value}?
3. Create your first evaluator
Next, we will create an evaluator that will assess the outputs of our LLM.
Navigate to Add Evaluator and choose LLM-As-A-Judge
From the evaluation templates choose Human vs AI
Adjust the variables in the template to match the columns of this dataset
[BEGIN DATA]
************
[Question]: Who invented {attributes.output.value}?
************
[Human Ground Truth Answer]: {attributes.input.value}
************
[AI Answer]: {output}
************
[END DATA]Finish by creating the evaluator
4. Run the task and evaluator
Finally, you can run the task in the playground. Navigate to the experiment to see the outputs and evaluation results.
Last updated
Was this helpful?

