Menu
Open source

prometheus.exporter.kafka

The prometheus.exporter.kafka component embeds kafka_exporter for collecting metrics from a kafka server.

Usage

alloy
prometheus.exporter.kafka "LABEL" {
    kafka_uris = KAFKA_URI_LIST
}

Arguments

You can use the following arguments to configure the exporter’s behavior. Omitted fields take their default values.

NameTypeDescriptionDefaultRequired
kafka_urisarray(string)Address array (host:port) of Kafka server.yes
instancestringTheinstancelabel for metrics, default is the hostname:port of the first kafka_uris. You must manually provide the instance value if there is more than one string in kafka_uris.no
use_saslboolConnect using SASL/PLAIN.no
use_sasl_handshakeboolOnly set this to false if using a non-Kafka SASL proxy.falseno
sasl_usernamestringSASL user name.no
sasl_passwordstringSASL user password.no
sasl_mechanismstringThe SASL SCRAM SHA algorithm sha256 or sha512 as mechanism.no
use_tlsboolConnect using TLS.no
ca_filestringThe optional certificate authority file for TLS client authentication.no
cert_filestringThe optional certificate file for TLS client authentication.no
key_filestringThe optional key file for TLS client authentication.no
insecure_skip_verifyboolIf set to true, the server’s certificate will not be checked for validity. This makes your HTTPS connections insecure.no
kafka_versionstringKafka broker version.2.0.0no
use_zookeeper_lagboolIf set to true, use a group from zookeeper.no
zookeeper_urisarray(string)Address array (hosts) of zookeeper server.no
kafka_cluster_namestringKafka cluster name.no
metadata_refresh_intervaldurationMetadata refresh interval.1mno
allow_concurrencyboolIf set to true, all scrapes trigger Kafka operations. Otherwise, they will share results. WARNING: Disable this on large clusters.trueno
max_offsetsintThe maximum number of offsets to store in the interpolation table for a partition.1000no
prune_interval_secondsintHow frequently should the interpolation table be pruned, in seconds.30no
topics_filter_regexstringRegex filter for topics to be monitored..*no
groups_filter_regexstringRegex filter for consumer groups to be monitored..*no

Exported fields

The following fields are exported and can be referenced by other components.

NameTypeDescription
targetslist(map(string))The targets that can be used to collect exporter metrics.

For example, the targets can either be passed to a discovery.relabel component to rewrite the targets’ label sets or to a prometheus.scrape component that collects the exposed metrics.

The exported targets use the configured in-memory traffic address specified by the run command.

Component health

prometheus.exporter.kafka is only reported as unhealthy if given an invalid configuration. In those cases, exported fields retain their last healthy values.

Debug information

prometheus.exporter.kafka does not expose any component-specific debug information.

Debug metrics

prometheus.exporter.kafka does not expose any component-specific debug metrics.

Example

This example uses a prometheus.scrape component to collect metrics from prometheus.exporter.kafka:

alloy
prometheus.exporter.kafka "example" {
  kafka_uris = ["localhost:9200"]
}

// Configure a prometheus.scrape component to send metrics to.
prometheus.scrape "demo" {
  targets    = prometheus.exporter.kafka.example.targets
  forward_to = [prometheus.remote_write.demo.receiver]
}

prometheus.remote_write "demo" {
  endpoint {
    url = PROMETHEUS_REMOTE_WRITE_URL

    basic_auth {
      username = USERNAME
      password = PASSWORD
    }
  }
}

Replace the following:

  • PROMETHEUS_REMOTE_WRITE_URL: The URL of the Prometheus remote_write-compatible server to send metrics to.
  • USERNAME: The username to use for authentication to the remote_write API.
  • PASSWORD: The password to use for authentication to the remote_write API.

Compatible components

prometheus.exporter.kafka has exports that can be consumed by the following components:

Note

Connecting some components may not be sensible or components may require further configuration to make the connection work correctly. Refer to the linked documentation for more details.