Caution
Grafana Alloy is the new name for our distribution of the OTel collector. Grafana Agent has been deprecated and is in Long-Term Support (LTS) through October 31, 2025. Grafana Agent will reach an End-of-Life (EOL) on November 1, 2025. Read more about why we recommend migrating to Grafana Alloy.
Important: This documentation is about an older version. It's relevant only to the release noted, many of the features and functions have been updated or replaced. Please view the current version.
prometheus.exporter.kafka
The prometheus.exporter.kafka
component embeds
kafka_exporter for collecting metrics from a kafka server.
Usage
prometheus.exporter.kafka "LABEL" {
kafka_uris = KAFKA_URI_LIST
}
Arguments
You can use the following arguments to configure the exporter’s behavior. Omitted fields take their default values.
Name | Type | Description | Default | Required |
---|---|---|---|---|
kafka_uris | array(string) | Address array (host:port) of Kafka server. | yes | |
instance | string | Theinstance label for metrics, default is the hostname:port of the first kafka_uris. You must manually provide the instance value if there is more than one string in kafka_uris. | no | |
use_sasl | bool | Connect using SASL/PLAIN. | no | |
use_sasl_handshake | bool | Only set this to false if using a non-Kafka SASL proxy. | false | no |
sasl_username | string | SASL user name. | no | |
sasl_password | string | SASL user password. | no | |
sasl_mechanism | string | The SASL SCRAM SHA algorithm sha256 or sha512 as mechanism. | no | |
use_tls | bool | Connect using TLS. | no | |
ca_file | string | The optional certificate authority file for TLS client authentication. | no | |
cert_file | string | The optional certificate file for TLS client authentication. | no | |
key_file | string | The optional key file for TLS client authentication. | no | |
insecure_skip_verify | bool | If set to true, the server’s certificate will not be checked for validity. This makes your HTTPS connections insecure. | no | |
kafka_version | string | Kafka broker version. | 2.0.0 | no |
use_zookeeper_lag | bool | If set to true, use a group from zookeeper. | no | |
zookeeper_uris | array(string) | Address array (hosts) of zookeeper server. | no | |
kafka_cluster_name | string | Kafka cluster name. | no | |
metadata_refresh_interval | duration | Metadata refresh interval. | 1m | no |
allow_concurrency | bool | If set to true, all scrapes trigger Kafka operations. Otherwise, they will share results. WARNING: Disable this on large clusters. | true | no |
max_offsets | int | The maximum number of offsets to store in the interpolation table for a partition. | 1000 | no |
prune_interval_seconds | int | How frequently should the interpolation table be pruned, in seconds. | 30 | no |
topics_filter_regex | string | Regex filter for topics to be monitored. | .* | no |
groups_filter_regex | string | Regex filter for consumer groups to be monitored. | .* | no |
Exported fields
The following fields are exported and can be referenced by other components.
Name | Type | Description |
---|---|---|
targets | list(map(string)) | The targets that can be used to collect exporter metrics. |
For example, the targets
can either be passed to a discovery.relabel
component to rewrite the targets’ label sets, or to a prometheus.scrape
component that collects the exposed metrics.
The exported targets will use the configured in-memory traffic address specified by the run command.
Component health
prometheus.exporter.kafka
is only reported as unhealthy if given
an invalid configuration. In those cases, exported fields retain their last
healthy values.
Debug information
prometheus.exporter.kafka
does not expose any component-specific
debug information.
Debug metrics
prometheus.exporter.kafka
does not expose any component-specific
debug metrics.
Example
This example uses a prometheus.scrape
component to collect metrics
from prometheus.exporter.kafka
:
prometheus.exporter.kafka "example" {
kafka_uris = ["localhost:9200"]
}
// Configure a prometheus.scrape component to send metrics to.
prometheus.scrape "demo" {
targets = prometheus.exporter.kafka.example.targets
forward_to = [prometheus.remote_write.demo.receiver]
}
prometheus.remote_write "demo" {
endpoint {
url = PROMETHEUS_REMOTE_WRITE_URL
basic_auth {
username = USERNAME
password = PASSWORD
}
}
}
Replace the following:
PROMETHEUS_REMOTE_WRITE_URL
: The URL of the Prometheus remote_write-compatible server to send metrics to.USERNAME
: The username to use for authentication to the remote_write API.PASSWORD
: The password to use for authentication to the remote_write API.