このページは日本語には対応しておりません。随時翻訳に取り組んでいます。翻訳に関してご質問やご意見ございましたら、お気軽にご連絡ください。

Overview

As your infrastructure and applications grow, so does your log volume and the complexity of the data. A large volume of logs can introduce a lot of noise and make it difficult to analyze and troubleshoot logs. Use Observability Pipelines’s processors to decide which logs are valuable and which ones are noisy and uninteresting, before sending your logs to their destinations. You can use the following processors in the Observability Pipeline Worker to manage your logs:

  • Filter: Add a query to send only a subset of logs based on your conditions.
  • Sample: Define a sampling rate to send only a subset of your logs.
  • Quota: Enforce daily limits on either the volume of log data or the number of log events.
  • Dedupe: Drop duplicate copies of your logs, for example, due to retries because of network issues.
  • Remap: Add, drop, or rename a field in your logs.

Select a log source to get started: