Data Observability Overview

이 페이지는 아직 한국어로 제공되지 않습니다. 번역 작업 중입니다.
현재 번역 프로젝트에 대한 질문이나 피드백이 있으신 경우 언제든지 연락주시기 바랍니다.

Overview

Data Observability (DO) helps data teams improve the reliability of data for analytics and AI applications and optimize the performance and costs of data pipelines. By unifying quality and jobs monitoring from production to consumption, teams can detect and remediate issues faster while optimizing cost and performance.

Datadog Data Observability end-to-end lineage with Spark job traces.

Key capabilities

  • Detect failures early: Catch bad data in warehouses like Snowflake, Databricks, and BigQuery through ML-powered monitors before dashboards, stakeholders, or AI models are impacted. Detect upstream pipeline failures in jobs run on Databricks, Spark, Airflow, or dbt.
  • Accelerate remediation: Triage faster using end-to-end lineage to pinpoint root causes, assess incident blast radius, and route to the right owner. View which job in the pipeline failed or was delayed, and pivot into job execution traces and logs to determine why.
  • Optimize cost & performance: Get visibility into the cost and efficiency of Spark and Databricks jobs and clusters, and use recommendations to optimize cluster configuration, code, and queries.
  • Unify end-to-end observability: Correlate data quality, pipeline execution, and infrastructure signals in one place, spanning the entire data lifecycle.

Get started