Phoenix
CtrlK
TypeScript APIPython APICommunityGitHubPhoenix Cloud
  • Documentation
  • Self-Hosting
  • SDK and API Reference
  • Integrations
  • Cookbooks
  • Learn
  • Release Notes
  • Featured Tutorials
  • Agent Cookbooks
  • Agent Demos
  • Tracing
    • Cookbooks
    • Structured Data Extraction
  • Prompt Engineering
    • Few Shot Prompting
    • ReAct Prompting
    • Chain-of-Thought Prompting
    • Prompt Optimization
    • LLM as a Judge Prompt Optimization
  • Datasets & Experiments
    • Cookbooks
    • Summarization
    • Text2SQL
  • Evaluation
    • Cookbooks
    • Evaluate RAG
    • Evaluate an Agent
    • OpenAI Agents SDK Cookbook
  • Retrieval & Inferences
    • Cookbooks
    • Embeddings Analysis
Powered by GitBook
On this page

Was this helpful?

Edit on GitHub
Export as PDF
  1. Datasets & Experiments

Cookbooks

Iteratively improve your LLM task by building datasets, running experiments, and evaluating performance using code and LLM-as-a-judge.

Cover

Comprehensive Use Cases

Text2SQL

Summarization Service

Email Text Extraction

Pairwise Evaluator

Cover

RAG Use Cases

Answer and Context Relevancy Evals

Response Guideline Evals

LlamaIndex RAG with Reranker

PreviousLLM as a Judge Prompt OptimizationNextSummarization

Last updated 3 months ago

Was this helpful?

Platform

  • Tracing
  • Prompts
  • Datasets and Experiments
  • Evals

Software

  • Python Client
  • TypeScript Client
  • Phoenix Evals
  • Phoenix Otel

Resources

  • Container Images
  • X
  • Blue Sky
  • Blog

Integrations

  • OpenTelemetry
  • AI Providers

© 2025 Arize AI