Información general

Métricas de OpenTelemetry Apache Spark en un dashboard de Spark

El receptor de Apache Spark permite recopilar métricas de Apache Spark y acceder al dashboard de Información general de Spark. Configura el receptor según las especificaciones de la última versión del apachesparkreceiver.

Para más información, consulta la documentación del proyecto de OpenTelemetry para el receptor de Apache Spark.

Configuración

Para recopilar métricas de Apache Spark con OpenTelemetry para su uso con Datadog:

  1. Configura el receptor de Apache Spark en tu configuración de OpenTelemetry Collector.
  2. Asegúrate de que el OpenTelemetry Collector está configurado para exportar a Datadog.

Consulta la documentación del receptor de Apache Spark para obtener información detallada sobre las opciones y requisitos de configuración.

Datos recopilados

OTELDATADOGDESCRIPTIONFILTERTRANSFORM
spark.driver.block_manager.disk.usagespark.driver.disk_usedDisk space used by the BlockManager.× 1048576
spark.driver.block_manager.memory.usagespark.driver.memory_usedMemory usage for the driver’s BlockManager.× 1048576
spark.driver.dag_scheduler.stage.countspark.stage.countNumber of stages the DAGScheduler is either running or needs to run.
spark.executor.disk.usagespark.executor.disk_usedDisk space used by this executor for RDD storage.
spark.executor.disk.usagespark.rdd.disk_usedDisk space used by this executor for RDD storage.
spark.executor.memory.usagespark.executor.memory_usedStorage memory used by this executor.
spark.executor.memory.usagespark.rdd.memory_usedStorage memory used by this executor.
spark.job.stage.activespark.job.num_active_stagesNumber of active stages in this job.
spark.job.stage.resultspark.job.num_completed_stagesNumber of stages with a specific result in this job.job_result: completed
spark.job.stage.resultspark.job.num_failed_stagesNumber of stages with a specific result in this job.job_result: failed
spark.job.stage.resultspark.job.num_skipped_stagesNumber of stages with a specific result in this job.job_result: skipped
spark.job.task.activespark.job.num_tasks{status: running}Number of active tasks in this job.
spark.job.task.resultspark.job.num_skipped_tasksNumber of tasks with a specific result in this job.job_result: skipped
spark.job.task.resultspark.job.num_failed_tasksNumber of tasks with a specific result in this job.job_result: failed
spark.job.task.resultspark.job.num_completed_tasksNumber of tasks with a specific result in this job.job_result: completed
spark.stage.io.recordsspark.stage.input_recordsNumber of records written and read in this stage.direction: in
spark.stage.io.recordsspark.stage.output_recordsNumber of records written and read in this stage.direction: out
spark.stage.io.sizespark.stage.input_bytesAmount of data written and read at this stage.direction: in
spark.stage.io.sizespark.stage.output_bytesAmount of data written and read at this stage.direction: out
spark.stage.shuffle.io.read.sizespark.stage.shuffle_read_bytesAmount of data read in shuffle operations in this stage.
spark.stage.shuffle.io.recordsspark.stage.shuffle_read_recordsNumber of records written or read in shuffle operations in this stage.direction: in
spark.stage.shuffle.io.recordsspark.stage.shuffle_write_recordsNumber of records written or read in shuffle operations in this stage.direction: out

Consulta Asignación de métricas de OpenTelemetry para obtener más información.

Referencias adicionales

Más enlaces, artículos y documentación útiles: