Track disk utilization and failed volumes on each of your HDFS DataNodes. This Agent check collects metrics for these, as well as block- and cache-related metrics.
Use this check (hdfs_datanode) and its counterpart check (hdfs_namenode), not the older two-in-one check (hdfs); that check is deprecated.
Setup
Follow the instructions below to install and configure this check for an Agent running on a host. For containerized environments, see the Autodiscovery Integration Templates for guidance on applying these instructions.
Installation
The HDFS DataNode check is included in the Datadog Agent package, so you don’t need to install anything else on your DataNodes.
Configuration
Connect the Agent
Host
To configure this check for an Agent running on a host:
init_config:instances:## @param hdfs_datanode_jmx_uri - string - required## The HDFS DataNode check retrieves metrics from the HDFS DataNode's JMX## interface via HTTP(S) (not a JMX remote connection). This check must be installed on a HDFS DataNode. The HDFS## DataNode JMX URI is composed of the DataNode's hostname and port.#### The hostname and port can be found in the hdfs-site.xml conf file under## the property dfs.datanode.http.address## https://hadoop.apache.org/docs/r3.1.3/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml#- hdfs_datanode_jmx_uri:http://localhost:9864