CrewAI Tracing

Instrument multi-agent applications using CrewAI

Launch Phoenix

Install

CrewAI uses either Langchain or LiteLLM under the hood to call models, depending on the version.

If you're using CrewAI<0.63.0, we recommend installing our openinference-instrumentation-langchain library to get visibility of LLM calls.

If you're using CrewAI>= 0.63.0, we recommend instead adding our openinference-instrumentation-litellm library to get visibility of LLM calls.

Setup

Connect to your Phoenix instance using the register function.

Run CrewAI

From here, you can run CrewAI as normal

Observe

Now that you have tracing setup, all calls to your Crew will be streamed to your running Phoenix for observability and evaluation.

Resources

Last updated

Was this helpful?