- 필수 기능
- 시작하기
- Glossary
- 표준 속성
- Guides
- Agent
- 통합
- 개방형텔레메트리
- 개발자
- Administrator's Guide
- API
- Datadog Mobile App
- CoScreen
- Cloudcraft
- 앱 내
- 서비스 관리
- 인프라스트럭처
- 애플리케이션 성능
- APM
- Continuous Profiler
- 스팬 시각화
- 데이터 스트림 모니터링
- 데이터 작업 모니터링
- 디지털 경험
- 소프트웨어 제공
- 보안
- AI Observability
- 로그 관리
- 관리
Sensitive Data Scanner in the Cloud scans telemetry data, such as your application logs, APM events, RUM events, and events from Event Management. The data that can be scanned and redacted are:
You submit logs and events to the Datadog backend, so the data leaves your environment before it gets redacted. The logs and events are scanned and redacted in the Datadog backend during processing, so sensitive data is redacted before events are indexed and shown in the Datadog UI.
If you don’t want data to leave your environment before it gets redacted, use Observability Pipelines and the Sensitive Data Scanner processor to scan and redact sensitive data. See Set Up Pipelines for information on how to set up a pipeline and its components.
To use Sensitive Data Scanner in the Cloud, set up a scanning group to define what data to scan and then add scanning rules to determine what sensitive information to match within the data.
This document goes through the following:
By default, users with the Datadog Admin role have access to view and set up scanning rules. To allow other users access, grant the data_scanner_read
or data_scanner_write
permissions under Compliance to a custom role. See Access Control for details on how to set up roles and permissions.
A scanning group determines what data to scan. It consists of a query filter and a set of toggles to enable scanning for logs, APM, RUM, and events. See the Log Search Syntax documentation to learn more about query filters.
For Terraform, see the Datadog Sensitive Data Scanner group resource.
To set up a scanning group, perform the following steps:
By default, a newly-created scanning group is disabled. To enable a scanning group, click the corresponding toggle on the right side.
A scanning rule determines what sensitive information to match within the data defined by a scanning group. You can add predefined scanning rules from Datadog’s Scanning Rule Library or create your own rules using regex patterns. The data is scanned at ingestion time during processing. For logs, this means the scan is done before indexing and other routing decisions.
For Terraform, see the Datadog Sensitive Data Scanner rule resource.
To add scanning rules, perform the following steps:
Add scanning rule from the library rules
The Scanning Rule Library contains predefined rules for detecting common patterns such as email addresses, credit card numbers, API keys, authorization tokens, and more.
sensitive_data
및 sensitive_data_category
태그를 사용할 것을 권장합니다. 이러한 태그는 검색, 대시보드 및 모니터에서 사용할 수 있습니다. 민감한 데이터가 있는 로그 에 대한 액세스 제어](#control-access-to-로그-with-sensitive-data)를 사용하여 민감한 정보가 포함된 로그에 액세스할 수 있는 사용자를 결정하는 방법에 대한 자세한 내용은 태그를 참조하세요.After adding OOTB scanning rules, you can edit each rule separately and add additional keywords to the keyword dictionary.
Add a custom scanning rule
You can create custom scanning rules using regex patterns to scan for sensitive data.
\C
“single-byte” directive (which breaks UTF-8 sequences)\R
newline match\K
start of match reset directivevisa
, credit
, and card
. You can also require that these keywords be within a specified number of characters of a match. By default, keywords must be within 30 characters before a matched value.sensitive_data
및 sensitive_data_category
태그를 사용할 것을 권장합니다. 이러한 태그는 검색, 대시보드 및 모니터에서 사용할 수 있습니다. 민감한 데이터가 있는 로그 에 대한 액세스 제어](#control-access-to-로그-with-sensitive-data)를 사용하여 민감한 정보가 포함된 로그에 액세스할 수 있는 사용자를 결정하는 방법에 대한 자세한 내용은 태그를 참조하세요.Notes:
See Investigate Sensitive Data Issues for details on how to use the Summary page to triage your sensitive data issues.
There are reserved keywords that the Datadog platform requires for functionality. If any of these words are in a log that is being scanned, the 30 characters after the matched word are ignored and not redacted. For example, what comes after the word date
in a log is usually the event timestamp. If the timestamp is accidentally redacted, that would result in issues with processing the log and being able to query it later. Therefore, the behavior for excluded namespaces is to prevent unintentionally redacting important information for product functionality.
The excluded namespaces are:
host
hostname
syslog.hostname
service
status
env
dd.trace_id
trace_id
trace id
dd.span_id
span_id
span id
@timestamp
timestamp
_timestamp
Timestamp
date
published_date
syslog.timestamp
error.fingerprint
x-datadog-parent-id
visa
, credit
, and card
.sensitive_data
and sensitive_data_category
tags, which can be used in searches, dashboards, and monitors. See Control access to logs with sensitive data for information on how to use tags to determine who can access logs that contain sensitive data.To control who can access logs containing sensitive data, use tags added by the Sensitive Data Scanner to build queries with role-based access control (RBAC). You can restrict access to specific individuals or teams until the data ages out after the retention period. See How to Set Up RBAC for Logs for more information.
To redact sensitive data contained in tags, you must remap the tag to an attribute and then redact the attribute. Uncheck Preserve source attribute
in the remapper processor so that the tag is not preserved during the remapping.
To remap the tag to an attribute:
To redact the attribute:
To turn off Sensitive Data Scanner entirely, set the toggle to off for each Scanning Group so that they are disabled.