Grafana Cloud

Browse and debug conversations

The Conversations view in the AI Observability plugin lets you search, filter, and inspect full conversation threads. Use it to understand what happened during an agent interaction, identify failures, and analyze performance.

Search conversations

Navigate to Conversations in the AI Observability plugin. Use the search bar and filters to find conversations by:

  • Time range.
  • Agent name or version.
  • Model provider or name.
  • Conversation content.
  • Error status.
  • Tags or metadata values.
  • Evaluation results (pass, fail, or specific evaluator scores).

Conversations display a summary showing the agent, model, token count, cost, duration, and any evaluation scores. The metrics bar shows evaluation pass, fail, and neutral counts for each conversation.

Inspect a conversation

Click a conversation to open the detail view. The conversation detail shows:

  • A timeline of all generations in the conversation.
  • Each generation’s input messages, output messages, tool calls, and tool results.
  • Token usage breakdown (input, output, cache read, cache creation, reasoning).
  • Timing data (start, first token, end).
  • Evaluation scores attached to each generation.

Tool calls within a generation are grouped together, and thinking/reasoning content is collapsible. Generations with no captured content show an inline placeholder.

Click the error badge in the metrics bar to jump directly to the first errored generation in the conversation.

View the dependency graph

For multi-agent pipelines where generations declare dependencies using depends_on, the conversation detail includes a Graph tab that visualizes the dependency DAG. Each node represents a generation, and edges show parent-child relationships. Use this view to understand how agents in a pipeline interact and to trace quality signals across dependencies.

Trace drilldown

Each generation links to its OpenTelemetry trace. Click the trace link to open the trace view in Tempo, where you can see the full span tree including:

  • LLM call duration and error status.
  • Tool execution spans.
  • Nested agent invocations.

Use trace drilldown to identify latency bottlenecks, understand tool call sequences, and diagnose timeout or error scenarios.

Provide feedback

Annotate conversations

Open the Annotations panel from the conversation detail view to add threaded notes. Annotations support key-value tags and are visible to all team members with the Feedback Writer or Admin role. Use annotations to:

  • Flag conversations for review.
  • Add context about specific agent behaviors.
  • Track investigation notes alongside the conversation data.

The conversation list shows an annotation count badge, and you can browse annotated conversations from a dedicated pane in the conversation browser.

Next steps