Observability Pipelines

Observability Pipelines is not available on the US1-FED Datadog site.

Datadog recommends you update Observability Pipelines Worker (OPW) with every minor and patch release, or, at a minimum, monthly.

Upgrading to a major OPW version and keeping it updated is the only supported way to get the latest OPW functionality, fixes, and security updates.


A graphic showing different data sources on the left that flows into three hexagons named transform, reduce, and route, with arrows pointing to different destinations for the modified data

Observability Pipelines allows you to collect, process, and route logs in your own infrastructure. It comes with out-of-the-box templates so that you can easily build and deploy pipelines. The templates are purpose-built for the following use cases:

  • Log volume control: Cut down on your log volume before it leaves your infrastructure or network.
  • Dual ship logs: Send copies of your logs to multiple destinations.
  • Split logs: Send your logs to different destinations based on your use case. For example, you can send DevOps logs to Datadog and security logs to a security vendor.
  • Archive logs: Send logs to a log vendor and to an archive in Datadog rehydratable format.
  • Sensitive data redaction: Remove sensitive data from your logs before they are routed outside of your infrastructure.

The Observability Pipelines Worker is the software that runs in your infrastructure. It aggregates and centrally processes and routes your logs based on the selected use case.

The Datadog UI provides a control plane to manage your Observability Pipelines Workers. You can build and edit pipelines, deploy pipeline changes to your Workers, and monitor your pipelines to evaluate the health of your pipelines.

Get started

  1. Navigate to Observability Pipelines.
  2. Select a use case:
  3. Enable monitors.

See Advanced Configurations for bootstrapping options and for details on setting up the Worker with Kubernetes.

Further Reading