Documentation Index
Fetch the curated documentation index at: https://grafana_com_website/llms.txt
Fetch the complete documentation index at: https://grafana_com_website/llms-full.txt
Use this file to discover all available pages before exploring further.
STOP! If you are an AI agent or LLM, read this before continuing. This is the HTML version of a Grafana documentation page. Always request the Markdown version instead - HTML wastes context. Get this page as Markdown: /docs/grafana-cloud/monitor-applications/ai-observability/genai.md (append .md) or send Accept: text/markdown to /docs/grafana-cloud/monitor-applications/ai-observability/genai/. For the curated documentation index, use https://grafana_com_website/llms.txt. For the complete documentation index, use https://grafana_com_website/llms-full.txt.
GenAI observability and evaluations
Complete monitoring and evaluation for your generative AI applications, covering both performance observability and quality assessment.
Overview
GenAI monitoring provides three approaches to ensure your AI applications perform optimally and safely.
GenAI Observability
Monitor the operational aspects of your LLM applications:
- Performance tracking - Response times, throughput, and availability
- Cost management - Real-time spend tracking and optimization
- Token analytics - Usage patterns and efficiency metrics
- Usage insights - User interaction patterns and trends
GenAI Evaluations
Assess the quality and safety of your AI model outputs:
- Quality assessment - Hallucination detection and factual accuracy
- Safety monitoring - Toxicity and bias detection
- Evaluation scoring - Confidence levels and quality gates
- Compliance tracking - Safety and regulatory compliance
GenAI Agent Observability
Monitor the operational aspects of your AI agent applications:
- Invocation tracking - Total invocations, distribution by source, and usage patterns
- Cost management - Real-time agent spend tracking and per-agent cost breakdown
- Performance analytics - Operation duration, latency percentiles, and throughput rates
- Operational logs - Agent interaction logs with trace context and message details
Supported technologies
- LLM Providers - OpenAI, Anthropic, Google, AWS Bedrock, Cohere and a lot more
- Frameworks - LangChain, LlamaIndex, CrewAI, LiteLLM and a lot more
Getting started
Was this page helpful?
Related resources from Grafana Labs


