This is documentation for the next version of Grafana Alloy Documentation. For the latest stable release, go to the latest version.
prometheus.exporter.databricks
The prometheus.exporter.databricks component embeds the databricks_exporter for collecting billing, jobs, pipelines, and SQL warehouse metrics from Databricks System Tables via HTTP for Prometheus consumption.
Usage
prometheus.exporter.databricks "LABEL" {
server_hostname = "<DATABRICKS_SERVER_HOSTNAME>"
warehouse_http_path = "<DATABRICKS_WAREHOUSE_HTTP_PATH>"
client_id = "<DATABRICKS_CLIENT_ID>"
client_secret = "<DATABRICKS_CLIENT_SECRET>"
}Arguments
You can use the following arguments with prometheus.exporter.databricks:
Lookback windows
The exporter queries Databricks System Tables using SQL with sliding time windows. Each scrape collects data from now - lookback to now:
billing_lookback: Queriessystem.billing.usagefor DBU consumption and cost estimates. Databricks billing data typically has 24-48 hour lag.jobs_lookback: Queriessystem.lakeflow.job_run_timelinefor job run counts, durations, and status.pipelines_lookback: Queriessystem.lakeflow.pipeline_event_logfor DLT pipeline metrics.queries_lookback: Queriessystem.query.historyfor SQL warehouse query metrics.
The lookback window should be at least 2x the scrape_interval to ensure data continuity between scrapes. For example, with a 10-minute scrape interval, use at least 20 minutes of lookback.
Blocks
The prometheus.exporter.databricks component doesn’t support any blocks. You can configure this component with arguments.
Exported fields
The following fields are exported and can be referenced by other components.
For example, the targets can either be passed to a discovery.relabel component to rewrite the targets’ label sets or to a prometheus.scrape component that collects the exposed metrics.
The exported targets use the configured in-memory traffic address specified by the run command.
Component health
prometheus.exporter.databricks is only reported as unhealthy if given an invalid configuration.
In those cases, exported fields retain their last healthy values.
Debug information
prometheus.exporter.databricks doesn’t expose any component-specific debug information.
Debug metrics
prometheus.exporter.databricks doesn’t expose any component-specific debug metrics.
Prerequisites
Before using this component, you need:
- Databricks Workspace with Unity Catalog and System Tables enabled
- Service Principal with OAuth2 M2M authentication configured
- SQL Warehouse for querying System Tables and serverless is recommended for cost efficiency
Refer to the Databricks documentation for detailed OAuth2 M2M setup instructions.
Example
The following example uses a prometheus.scrape component to collect metrics from prometheus.exporter.databricks:
prometheus.exporter.databricks "example" {
server_hostname = "dbc-abc123-def456.cloud.databricks.com"
warehouse_http_path = "/sql/1.0/warehouses/xyz789"
client_id = "my-service-principal-id"
client_secret = "my-service-principal-secret"
}
// Configure a prometheus.scrape component to collect databricks metrics.
prometheus.scrape "demo" {
targets = prometheus.exporter.databricks.example.targets
forward_to = [prometheus.remote_write.demo.receiver]
scrape_interval = "10m"
scrape_timeout = "9m"
}
prometheus.remote_write "demo" {
endpoint {
url = "<PROMETHEUS_REMOTE_WRITE_URL>"
basic_auth {
username = "<USERNAME>"
password = "<PASSWORD>"
}
}
}Replace the following:
<PROMETHEUS_REMOTE_WRITE_URL>: The URL of the Prometheusremote_writecompatible server to send metrics to.<USERNAME>: The username to use for authentication to theremote_writeAPI.<PASSWORD>: The password to use for authentication to theremote_writeAPI.
Tuning recommendations
scrape_interval: Use 10-30 minutes. The exporter queries Databricks System Tables which can be slow. Increase thescrape_intervalto reduce your SQL Warehouse costs.scrape_timeout: Must be less thanscrape_interval. The exporter typically takes 90-120 seconds per scrape depending on data volume.- Lookback vs interval: The lookback windows should be at least 2x the scrape interval. The defaults,
3hfor jobs and pipelines, and2hfor queries, work well with 10-30 minute scrape intervals.
High cardinality warning
The collect_task_retries flag adds task-level retry metrics which can significantly increase cardinality for workspaces with many jobs.
Only enable this feature if you really need it.
Compatible components
prometheus.exporter.databricks has exports that can be consumed by the following components:
- Components that consume Targets
Note
Connecting some components may not be sensible or components may require further configuration to make the connection work correctly. Refer to the linked documentation for more details.



