Book a Demo!
CoCalc Logo Icon
StoreFeaturesDocsShareSupportNewsAboutPoliciesSign UpSign In
aos
GitHub Repository: aos/grafana-agent
Path: blob/main/docs/sources/static/configuration/integrations/kafka-exporter-config.md
4096 views
---
title: kafka_exporter_config aliases: - ../../../configuration/integrations/kafka-exporter-config/
---

kafka_exporter_config

The kafka_exporter_config block configures the kafka_exporter integration, which is an embedded version of kafka_exporter. This allows for the collection of Kafka Lag metrics and exposing them as Prometheus metrics.

We strongly recommend that you configure a separate user for the Agent, and give it only the strictly mandatory security privileges necessary for monitoring your node, as per the documentation.

Full reference of options:

# Enables the kafka_exporter integration, allowing the Agent to automatically # collect system metrics from the configured dnsmasq server address [enabled: <boolean> | default = false] # Sets an explicit value for the instance label when the integration is # self-scraped. Overrides inferred values. # # The default value for this integration is inferred from the hostname # portion of the first kafka_uri value. If there is more than one string # in kafka_uri, the integration will fail to load and an instance value # must be manually provided. [instance: <string>] # Automatically collect metrics from this integration. If disabled, # the dnsmasq_exporter integration will be run but not scraped and thus not # remote-written. Metrics for the integration will be exposed at # /integrations/dnsmasq_exporter/metrics and can be scraped by an external # process. [scrape_integration: <boolean> | default = <integrations_config.scrape_integrations>] # How often should the metrics be collected? Defaults to # prometheus.global.scrape_interval. [scrape_interval: <duration> | default = <global_config.scrape_interval>] # The timeout before considering the scrape a failure. Defaults to # prometheus.global.scrape_timeout. [scrape_timeout: <duration> | default = <global_config.scrape_timeout>] # Allows for relabeling labels on the target. relabel_configs: [- <relabel_config> ... ] # Relabel metrics coming from the integration, allowing to drop series # from the integration that you don't care about. metric_relabel_configs: [ - <relabel_config> ... ] # How frequent to truncate the WAL for this integration. [wal_truncate_frequency: <duration> | default = "60m"] # Address array (host:port) of Kafka server [kafka_uris: <[]string>] # Connect using SASL/PLAIN [use_sasl: <bool>] # Only set this to false if using a non-Kafka SASL proxy [use_sasl_handshake: <bool> | default = true] # SASL user name [sasl_username: <string>] # SASL user password [sasl_password: <string>] # The SASL SCRAM SHA algorithm sha256 or sha512 as mechanism [sasl_mechanism: <string>] # Connect using TLS [use_tls: <bool>] # The optional certificate authority file for TLS client authentication [ca_file: <string>] # The optional certificate file for TLS client authentication [cert_file: <string>] # The optional key file for TLS client authentication [key_file: <string>] # If true, the server's certificate will not be checked for validity. This will make your HTTPS connections insecure [insecure_skip_verify: <bool>] # Kafka broker version [kafka_version: <string> | default = "2.0.0"] # if you need to use a group from zookeeper [use_zookeeper_lag: <bool>] # Address array (hosts) of zookeeper server. [zookeeper_uris: <[]string>] # Kafka cluster name [kafka_cluster_name: <string>] # Metadata refresh interval [metadata_refresh_interval: <duration> | default = "1m"] # If true, all scrapes will trigger kafka operations otherwise, they will share results. WARN: This should be disabled on large clusters [allow_concurrency: <bool> | default = true] # Maximum number of offsets to store in the interpolation table for a partition [max_offsets: <int> | default = 1000] # How frequently should the interpolation table be pruned, in seconds [prune_interval_seconds: <int> | default = 30] # Regex filter for topics to be monitored [topics_filter_regex: <string> | default = ".*"] # Regex filter for consumer groups to be monitored [groups_filter_regex: <string> | default = ".*"]