이 페이지는 아직 한국어로 제공되지 않으며 번역 작업 중입니다. 번역에 관한 질문이나 의견이 있으시면 언제든지 저희에게 연락해 주십시오.
Data Streams Monitoring is not available for the site.
Data Streams Monitoring provides a standardized method for teams to understand and manage pipelines at scale by making it easy to:
- Measure pipeline health with end-to-end latencies for events traversing across your system.
- Pinpoint faulty producers, consumers or queues, then pivot to related logs or clusters to troubleshoot faster.
- Prevent cascading delays by equipping service owners to stop backed up events from overwhelming downstream services.
To get started, follow the installation instructions to configure services with Data Streams Monitoring:
|Java/Scala||Kafka (self-hosted, Amazon MSK, Confluent Cloud / Platform), RabbitMQ, HTTP, gRPC, Amazon SQS|
|Python||Kafka (self-hosted, Amazon MSK, Confluent Cloud / Platform), Amazon SQS, Amazon Kinesis|
|.NET||Kafka (self-hosted, Amazon MSK, Confluent Cloud / Platform), RabbitMQ|
|Node.js||Kafka (self-hosted, Amazon MSK, Confluent Cloud / Platform)|
|Go||All (with manual instrumentation)|
Explore Data Streams Monitoring
Measure end-to-end pipeline health with new metrics
Once Data Streams Monitoring is configured, you can measure the time it usually takes for events to traverse between any two points in your asynchronous system:
|Metric Name||Notable Tags||Description|
|End to end latency of a pathway from a specified source to destination service|
|Lag in seconds between producer and consumer. Requires Java Agent v1.9.0 or later.|
You can also graph and visualize these metrics on any dashboard or notebook:
Monitor end-to-end latency of any pathway
Depending on how events traverse through your system, different paths can lead to increased latency. With the Pathways tab, you can view latency between any two points throughout your pipelines, including queues, producers, and consumers to identify bottlenecks and optimize performance. Easily create a monitor for a pathway, or export to a dashboard.
Attribute incoming messages to any queue, service, or cluster
High lag on a consuming service, increased resource use on a Kafka broker, and increased RabbitMQ or Amazon SQS queue size are frequently explained by changes in the way adjacent services are producing to or consuming from these entities.
Click on the Throughput tab on any service or queue in Data Streams Monitoring to quickly detect changes in throughput, and which upstream or downstream service these changes originate from. Once the Service Catalog is configured, you can immediately pivot to the corresponding team’s Slack channel or on-call engineer.
By filtering to a single Kafka, RabbitMQ, or Amazon SQS cluster, you can detect changes in incoming or outgoing traffic for all detected topics or queues running on that cluster:
Quickly pivot to identify root causes in infrastructure, logs, or traces
Datadog automatically links the infrastructure powering your services and related logs through Unified Service Tagging, so you can easily localize bottlenecks. Click the Infra or Logs tabs to further troubleshoot why pathway latency or consumer lag has increased. To view traces within your pathways, click the Processing Latency tab.
Additional helpful documentation, links, and articles: