kafka_exporter_config
DocumentationGrafana CloudInstrument and send dataGrafana AgentStatic modeConfigure static modeintegrations_configkafka_exporter_config
Grafana Cloud
kafka_exporter_config
The kafka_exporter_config
block configures the kafka_exporter
integration, which is an embedded version of kafka_exporter
.
This allows for the collection of Kafka Lag metrics and exposing them as Prometheus metrics.
We strongly recommend that you configure a separate user for the Agent, and give it only the strictly mandatory security privileges necessary for monitoring your node, as per the documentation.
Full reference of options:
yaml
# Enables the kafka_exporter integration, allowing the Agent to automatically
# collect system metrics from the configured dnsmasq server address
[enabled: <boolean> | default = false]
# Sets an explicit value for the instance label when the integration is
# self-scraped. Overrides inferred values.
#
# The default value for this integration is inferred from the hostname
# portion of the first kafka_uri value. If there is more than one string
# in kafka_uri, the integration will fail to load and an instance value
# must be manually provided.
[instance: <string>]
# Automatically collect metrics from this integration. If disabled,
# the dnsmasq_exporter integration will be run but not scraped and thus not
# remote-written. Metrics for the integration will be exposed at
# /integrations/dnsmasq_exporter/metrics and can be scraped by an external
# process.
[scrape_integration: <boolean> | default = <integrations_config.scrape_integrations>]
# How often should the metrics be collected? Defaults to
# prometheus.global.scrape_interval.
[scrape_interval: <duration> | default = <global_config.scrape_interval>]
# The timeout before considering the scrape a failure. Defaults to
# prometheus.global.scrape_timeout.
[scrape_timeout: <duration> | default = <global_config.scrape_timeout>]
# Allows for relabeling labels on the target.
relabel_configs:
[- <relabel_config> ... ]
# Relabel metrics coming from the integration, allowing to drop series
# from the integration that you don't care about.
metric_relabel_configs:
[ - <relabel_config> ... ]
# How frequent to truncate the WAL for this integration.
[wal_truncate_frequency: <duration> | default = "60m"]
# Address array (host:port) of Kafka server
[kafka_uris: <[]string>]
# Connect using SASL/PLAIN
[use_sasl: <bool>]
# Only set this to false if using a non-Kafka SASL proxy
[use_sasl_handshake: <bool> | default = true]
# SASL user name
[sasl_username: <string>]
# SASL user password
[sasl_password: <string>]
# The SASL SCRAM SHA algorithm sha256 or sha512 as mechanism
[sasl_mechanism: <string>]
# Configure the Kerberos client to not use PA_FX_FAST.
[sasl_disable_pafx_fast: <string>]
# Connect using TLS
[use_tls: <bool>]
# Used to verify the hostname on the returned certificates unless tls.insecure-skip-tls-verify is given. If you don't provide the Kafka server name, the hostname is taken from the URL.
[tls_server_name: <string>]
# The optional certificate authority file for TLS client authentication
[ca_file: <string>]
# The optional certificate file for TLS client authentication
[cert_file: <string>]
# The optional key file for TLS client authentication
[key_file: <string>]
# If true, the server's certificate will not be checked for validity. This will make your HTTPS connections insecure
[insecure_skip_verify: <bool>]
# Kafka broker version
[kafka_version: <string> | default = "2.0.0"]
# if you need to use a group from zookeeper
[use_zookeeper_lag: <bool>]
# Address array (hosts) of zookeeper server.
[zookeeper_uris: <[]string>]
# Kafka cluster name
[kafka_cluster_name: <string>]
# Metadata refresh interval
[metadata_refresh_interval: <duration> | default = "1m"]
# Service name when using kerberos Auth.
[gssapi_service_name: <string>]
# Kerberos config path.
[gssapi_kerberos_config_path: <string>]
# Kerberos realm.
[gssapi_realm: <string>]
# Kerberos keytab file path.
[gssapi_key_tab_path: <string>]
# Kerberos auth type. Either 'keytabAuth' or 'userAuth'.
[gssapi_kerberos_auth_type: <string>]
# Whether show the offset/lag for all consumer group, otherwise, only show connected consumer groups.
[offset_show_all: <bool> | default = true]
# Minimum number of topics to monitor.
[topic_workers: <int> | default = 100]
# If true, all scrapes will trigger kafka operations otherwise, they will share results. WARN: This should be disabled on large clusters
[allow_concurrency: <bool> | default = true]
# If true, the broker may auto-create topics that we requested which do not already exist.
[allow_auto_topic_creation: <bool>]
# Maximum number of offsets to store in the interpolation table for a partition
[max_offsets: <int> | default = 1000]
# Deprecated (no-op), use metadata_refresh_interval instead.
[prune_interval_seconds: <int> | default = 30]
# Regex filter for topics to be monitored
[topics_filter_regex: <string> | default = ".*"]
# Regex that determines which topics to exclude.
[topics_exclude_regex: <string> | default = "^$"]
# Regex filter for consumer groups to be monitored
[groups_filter_regex: <string> | default = ".*"]
# Regex that determines which consumer groups to exclude.
[groups_exclude_regex: <string> | default = "^$"]
Was this page helpful?
Related resources from Grafana Labs
Additional helpful documentation, links, and articles:
Video
Getting started with the Grafana LGTM Stack
In this webinar, we’ll demo how to get started using the LGTM Stack: Loki for logs, Grafana for visualization, Tempo for traces, and Mimir for metrics.
Video
Intro to Kubernetes monitoring in Grafana Cloud
In this webinar you’ll learn how Grafana offers developers and SREs a simple and quick-to-value solution for monitoring their Kubernetes infrastructure.
Video
Building advanced Grafana dashboards
In this webinar, we’ll demo how to build and format Grafana dashboards.