Menu
Grafana Cloud

prometheus.exporter.kafka

The prometheus.exporter.kafka component embeds kafka_exporter for collecting metrics from a kafka server.

Usage

alloy
prometheus.exporter.kafka "LABEL" {
    kafka_uris = KAFKA_URI_LIST
}

Arguments

You can use the following arguments to configure the exporter’s behavior. Omitted fields take their default values.

NameTypeDescriptionDefaultRequired
kafka_urisarray(string)Address array (host:port) of Kafka server.yes
instancestringTheinstancelabel for metrics, default is the hostname:port of the first kafka_uris. You must manually provide the instance value if there is more than one string in kafka_uris.no
use_saslboolConnect using SASL/PLAIN.no
use_sasl_handshakeboolOnly set this to false if using a non-Kafka SASL proxy.trueno
sasl_usernamestringSASL user name.no
sasl_passwordstringSASL user password.no
sasl_mechanismstringThe SASL SCRAM SHA algorithm sha256 or sha512 as mechanism.no
sasl_disable_pafx_fastboolConfigure the Kerberos client to not use PA_FX_FAST.no
use_tlsboolConnect using TLS.no
tls_server_namestringUsed to verify the hostname on the returned certificates unless tls.insecure-skip-tls-verify is given. If you don’t provide the Kafka server name, the hostname is taken from the URL.no
ca_filestringThe optional certificate authority file for TLS client authentication.no
cert_filestringThe optional certificate file for TLS client authentication.no
key_filestringThe optional key file for TLS client authentication.no
insecure_skip_verifyboolIf set to true, the server’s certificate will not be checked for validity. This makes your HTTPS connections insecure.no
kafka_versionstringKafka broker version.2.0.0no
use_zookeeper_lagboolIf set to true, use a group from zookeeper.no
zookeeper_urisarray(string)Address array (hosts) of zookeeper server.no
kafka_cluster_namestringKafka cluster name.no
metadata_refresh_intervaldurationMetadata refresh interval.1mno
gssapi_service_namestringService name when using Kerberos Authorizationno
gssapi_kerberos_config_pathstringKerberos config path.no
gssapi_realmstringKerberos realm.no
gssapi_key_tab_pathstringKerberos keytab file path.no
gssapi_kerberos_auth_typestringKerberos auth type. Either ‘keytabAuth’ or ‘userAuth’.no
offset_show_allboolIf true, the broker may auto-create topics that we requested which do not already exist.trueno
topic_workersintMinimum number of topics to monitor.100no
allow_concurrencyboolIf set to true, all scrapes trigger Kafka operations. Otherwise, they will share results. WARNING: Disable this on large clusters.trueno
allow_auto_topic_creationboolIf true, the broker may auto-create topics that we requested which do not already exist.no
max_offsetsintThe maximum number of offsets to store in the interpolation table for a partition.1000no
prune_interval_secondsintDeprecated (no-op), use metadata_refresh_interval instead.30no
topics_filter_regexstringRegex filter for topics to be monitored..*no
topics_exclude_regexstringRegex that determines which topics to exclude.^$no
groups_filter_regexstringRegex filter for consumer groups to be monitored..*no
groups_exclude_regexstringRegex that determines which consumer groups to exclude.^$no

Exported fields

The following fields are exported and can be referenced by other components.

NameTypeDescription
targetslist(map(string))The targets that can be used to collect exporter metrics.

For example, the targets can either be passed to a discovery.relabel component to rewrite the targets’ label sets or to a prometheus.scrape component that collects the exposed metrics.

The exported targets use the configured in-memory traffic address specified by the run command.

Component health

prometheus.exporter.kafka is only reported as unhealthy if given an invalid configuration. In those cases, exported fields retain their last healthy values.

Debug information

prometheus.exporter.kafka does not expose any component-specific debug information.

Debug metrics

prometheus.exporter.kafka does not expose any component-specific debug metrics.

Example

This example uses a prometheus.scrape component to collect metrics from prometheus.exporter.kafka:

alloy
prometheus.exporter.kafka "example" {
  kafka_uris = ["localhost:9200"]
}

// Configure a prometheus.scrape component to send metrics to.
prometheus.scrape "demo" {
  targets    = prometheus.exporter.kafka.example.targets
  forward_to = [prometheus.remote_write.demo.receiver]
}

prometheus.remote_write "demo" {
  endpoint {
    url = PROMETHEUS_REMOTE_WRITE_URL

    basic_auth {
      username = USERNAME
      password = PASSWORD
    }
  }
}

Replace the following:

  • PROMETHEUS_REMOTE_WRITE_URL: The URL of the Prometheus remote_write-compatible server to send metrics to.
  • USERNAME: The username to use for authentication to the remote_write API.
  • PASSWORD: The password to use for authentication to the remote_write API.

Compatible components

prometheus.exporter.kafka has exports that can be consumed by the following components:

Note

Connecting some components may not be sensible or components may require further configuration to make the connection work correctly. Refer to the linked documentation for more details.