New announcements for Serverless, Network, RUM, and more from Dash! New announcements from Dash!

Logging without Limits


Sometimes the amount of log events generated by your infrastructure is too large and fluctuates a lot, raising the issue of choosing which logs should be sent to a Log Management solution and which should be sent to an archive. But filtering your logs before sending them inevitably leads to gaps in coverage, and often filters out valuable data.

Datadog log management removes these limitations by decoupling log ingestion from indexing, which makes it possible to cost-effectively collect, process, and archive all your logs. Datadog Index Filters avoid complex Agent level configuration and control what you want to index dynamically.

With Logging without Limits*, you can now:

This flexibility is critical in some exceptional situations such as outages, when you can disable specific filters to send more data. The inverse is true as well; if you over-consume because of a seasonal reason (Black Friday, Christmas, etc…) you can decide to selectively reduce some volume to avoid overages.

Index details

Indexes are located in the pipeline page within the Indexes section. Double click on them or click on the edit button to see more information about the number of logs that were indexed in the past 3 days, and the retention period for those logs:

Indexed logs can be used for faceted searching, Log Analytics, dashboarding, and monitoring.

It is also possible to have multiple indexes with different retention periods (currently in private beta). Logs enter the first index whose filter they match on, so it is important to order your indexes carefully.

For example, if you create a first index filtered to the status:notice attribute and a second index filtered to the status:error attribute and a final one without any filter (the equivalent of *), all your notice logs would go to the first index, all your error logs to the second index and the rest would go to the final one.

Multiple indexes also provide the ability to define access rules on the data contained in each index. More information available in the role base access control documentation.

Setup Log Monitors on volumes

Get notified at any moment if the volumes in any scope (service, availibility-zone, etc…) of your infrastructure is growing unexpectedly:

  1. Go in the Datadog Log Explorer view
  2. Build a search query that represents the volume to monitor.
  3. Click on Export to monitor.
  4. Define the rate you would like to set as warning or error.
  5. Define an explicit notification: The volume on this service just got too high. Define an additional exclusion filter or increase the sampling rate to put it back under control.

Exclusion Filters

Index Filters give dynamic control over what goes into your indexes.

For example, if some logs were captured only for troubleshooting purposes, you may only care to index those logs with errors and warnings. This can easily be achieved with exclusion filters.

To define a new Index Filter click on the “add” button:

To configure an exclusion filter:

  1. Define the name of your filter
  2. Define the query for logs to exclude from your index Note: It is possible to use any attribute or tag in the Index filter query, even those that are not facets. If you are filtering by non-faceted attributes or tags, be sure to hit “enter/return” from the query bar
  3. Define the sampling rate
  4. Save the filter

Note: If a log matches several exclusion filters, only the first exclusion filter rule is applied. A log is not sampled or excluded multiple times by different exclusion filters.


The following filter removes all logs that have a fast response time. We use the duration attribute and filter all logs that have a value below 100ms.

    "http": {
        "url": "",
        "status_code": "200"
    "metadata": {
        "version": 12,
        "release": "sept18"

Filter: @duration:<100

Container example

Container logs have a lot of metadata collected as tags. To exclude all logs coming from images that contains httpd in the image_name tag use the following filter:

Filter: image_name:*httpd*

Reorder filters

Order matters for exclusion filters. And contrary to where several pipelines can process a log, if a log matches several exclusion filters, only the first exclusion filter rule is applied.

Reorder your pipeline to make sure the proper exclusion filters applies for your log. For instance, you probably want to set up the filters with the most inclusive queries after the others.

To reorder your exclusion filter, drag and drop them into your preferred order.

Enable/Disable filters

If all logs are not worth indexing on a daily basis, they might still be critical in some situations. Debug logs, for instance, are not always useful but during a complex troubleshooting or during a production release they become very interesting to get better insight into what is going on.

Instead of changing your application logging level or using a complex internal filtering tool, it is now possible to change what is indexed directly with the Datadog index filters.

Enable or disable them in one click in the Pipeline page:

Further Reading

*Logging without Limits is a trademark of Datadog, Inc.