Tracing is a powerful tool for understanding the behavior of your LLM application. Leveraging LLM tracing with Arize, you can track down issues around application latency, token usage, runtime exceptions, retrieved documents, embeddings, LLM parameters, prompt templates, tool descriptions, LLM function calls, and more. To get started, you can automatically collect traces from major frameworks and libraries using auto instrumentation from Arize — including for OpenAI, LlamaIndex, Mistral AI, DSPy, AWS Bedrock, and Autogen — or create and customize spans using the OpenTelemetry Trace API (because Arize and OpenInference support OpenTelemetry, this means that you can perform manual instrumentation with no LLM framework required).
This tutorial covers what LLM tracing is, why it matters, how to ingest traces, and LLM tracing in Arize. It includes a code-along example.
More:
🚀 Quickstart
🤖 Auto instrumentation
🧠 Learn more about LLM tracing
🪽 LLM Tracing with Phoenix