Datadog Application Performance Monitoring (APM) gives deep visibility into your applications with out-of-the-box performance dashboards for web services, queues, and databases to monitor requests, errors, and latency. Distributed traces seamlessly correlate to browser sessions, logs, profiles, synthetic checks, network, processes, and infrastructure metrics across hosts, containers, proxies, and serverless functions. Navigate directly from investigating a slow trace to identifying the specific line of code causing performance bottlenecks with code hotspots.
Control and manage data flowing into and being kept by Datadog
Traces start in your instrumented applications and flow into Datadog. For high-throughput services, you can view and control ingestion using Ingestion Controls. All ingested traces are available for live search and analytics for 15 minutes. You can use custom tag-based retention filters to keep exactly the traces that matter for your business for 15 days for search and analytics.
Generate custom metrics from spans
Generate metrics with 15-month retention from all ingested spans to create and monitor key business and performance indicators over time.
Connect traces with other telemetry
View your application logs side-by-side with the trace for a single distributed request with automatic trace-id injection. Link between real user sessions and traces to see the exact traces that correspond to user experiences and reported issues. Link simulated tests to traces to find the root cause of failures across frontend, network and backend requests.
Explore live and indexed traces
Search your ingested traces by any tag, live for 15 minutes. Analyze performance by any tag on any span during an outage to identify impacted users or transactions. View maps showing request flows and other visualizations to help you understand what your code is doing and where its performance can be improved.
Gain deep insight into your services
Understand service dependencies with an auto-generated service map from your traces alongside service performance metrics and monitor alert statuses.
Monitor Service metrics for requests, errors and latency percentiles. Analyze individual database queries or endpoints correlated with infrastructure.
Monitor service performance and compare between versions for rolling, blue/green, shadow, or canary deployments.
Profile your production code
Improve application latency and optimize compute resources with always-on production profiling to pinpoint the lines of code consuming the most CPU, memory, or I/O.
Further Reading
Additional helpful documentation, links, and articles: