1
Before running evals, make sure Phoenix is running & you have sent traces in your project. For more step by step instructions, check out this Get Started guide & Get Started with Tracing guide.
- Phoenix Cloud
- Local (Self-hosted)
Log in, create a space, navigate to the settings page in your space, and create your API keys.Set your environment variables.You can find your collector endpoint here:
Your Collector Endpoint is: https://app.phoenix.arize.com/s/ + your space name.

Launch your space, navigate to settings & copy your hostname for your collector endpoint
2
You’ll need to install the evals library that’s apart of Phoenix.
- Python
- TS
3
Since, we are running our evaluations on our trace data from our first project, we’ll need to pull that data into our code.
- Python
- TS
4
In this example, we will define, create, and run our own evaluator. There’s a number of different evaluators you can run, but this quick start will go through an LLM as a Judge Model.1) Define your LLM Judge ModelWe’ll use OpenAI as our evaluation model for this example, but Phoenix also supports a number of other models.If you haven’t yet defined your OpenAI API Key from the previous step, let’s first add it to our environment.
- Python
- TS
5
Now that we have defined our evaluator, we’re ready to evaluate our traces.
- Python
6
You’ll now be able to log your evaluations in your project view.
- Python
- TS

