Troubleshooting and Deep Dive for Kafka

Cette page n'est pas encore disponible en français, sa traduction est en cours.
Si vous avez des questions ou des retours sur notre projet de traduction actuel, n'hésitez pas à nous contacter.

Overview

“Kafka™ is used for building real-time data pipelines and streaming apps. It is horizontally scalable, fault-tolerant, wicked fast, and runs in production in thousands of companies.” - Official Kafka Site

Kafka is essentially a powerful, fast message brokering system used to transfer a payload/message from many applications to many applications. This is a Java based application that exposes metrics through mBeans.

Kafka components

There are four main components to Kafka:

  • Broker: Cluster of nodes responsible for establishing the mechanisms to write and read messages. (Main piece of Kafka, always in Java, usually managed by Apache Zookeeper)
  • Producer: Application(s) that is writing the messages that you are interested in viewing. (Most commonly in Java, but possibly in other languages)
  • Consumer: This is the application(s) that is receiving your set of messages. (Most commonly in Java, but possibly in other languages)
  • Topics - Mailboxes of messages that Producers and Consumers subscribe to. When writing or reading a message in Kafka, specify which “topic” you are to read from. You can think of this like a channel in slack, you join the ones you want to post and read messages to. Each topic then has a list of offsets that informs you where you are in the number of messages you have read/have left to read.

A more full dive into Kafka as well as on a Datadog Blogpost.

Datadog Kafka integrations

It is important to note that Datadog has two distinct Kafka Integrations. The first is named Kafka while the second is Kafka_Consumer.

The Kafka Integration uses Datadog’s JMXFetch application to pull metrics, just like our other Java based applications such as Cassandra, JMX, Tomcat, etc. This pulls metrics through the use of mBeans, where the engineering team has included a list of commonly used mBeans in the Kafka.yaml file. This can be extended with any other beans the user would like, or if your version of Kafka supports additional metrics.

The Kafka_Consumer Integration collects metrics like our standard Python based checks. This uses an internal Zookeeper API. Zookeeper is an Apache application that is responsible for managing the configuration for the cluster of nodes known as the Kafka broker. (In version 0.9 of Kafka things are a bit different, Zookeeper is no longer required, see the Troubleshooting section for more information). This check picks up only three metrics, and these do not come from JMXFetch.

Troubleshooting

Older Agent versions

This issue only applies if you are running version <5.20 of the Datadog Agent. In older versions of Kafka, consumer offsets were stored in Zookeper exclusively. The initial Kafka_consumer Agent Check was written when this limitation was in place. Due to this, you cannot get the kafka.consumer_lag metric if your offsets are stored in Kafka and you are using an older version of the Agent. Upgrade the Agent to the latest version to see these metrics.

Cannot connect to instance

You might see the following error for the Datadog-Kafka integration:

instance #kafka-localhost-<PORT_NUM> [ERROR]: 'Cannot connect to instance localhost:<PORT_NUM>. java.io.IOException: Failed to retrieve RMIServer stub

This error means the Datadog Agent is unable to connect to the Kafka instance to retrieve metrics from the exposed mBeans over the RMI protocol. The error can be resolved by including the following Java Virtual Machine (JVM) arguments when starting the Kafka instance (required for all separate Java instances - producer, consumer, and broker).

-Dcom.sun.management.jmxremote.port=<PORT_NUM> -Dcom.sun.management.jmxremote.rmi.port=<PORT_NUM>

Missing producer and consumer metrics

By default Datadog only collects broker based metrics.

For Java based producers and consumers, add the following to the conf.yaml and update the settings as necessary. See the sample kafka.d/conf.yaml for all available configuration options.

- host: remotehost
  port: 9998 # Producer
  tags:
    - kafka: producer0
- host: remotehost
  port: 9997 # Consumer
  tags:
    - kafka: consumer0

Note: This method does not work if you are using custom producer and consumer clients written in other languages or not exposing mBeans. To submit your metrics from your code, use DogStatsD.

Partition doesn’t exist

This issue is specifically for the Kafka Consumer Agent check. If you specify a partition in kafka_consumer.d/conf.yaml that doesn’t exist in your environment, you see the following error:

instance - #0 [Error]: ''

To remedy, specify the correct partition for your topic. This correlates to this line:

#     <TOPIC_NAME_1>: [0, 1, 4, 12]

Partition context limitation

The number of partition contexts collection is limited to 500. If you require more contexts, contact Datadog support.