---
title: Getting Started with Datadog
description: Datadog, the leading service for cloud-scale monitoring.
breadcrumbs: Docs > Infrastructure > Datadog Resource Catalog
---

# aws_personalize_data_deletion_job{% #aws_personalize_data_deletion_job %}

## `account_id`{% #account_id %}

**Type**: `STRING`

## `creation_date_time`{% #creation_date_time %}

**Type**: `TIMESTAMP`**Provider name**: `creationDateTime`**Description**: The creation date and time (in Unix time) of the data deletion job.

## `data_deletion_job_arn`{% #data_deletion_job_arn %}

**Type**: `STRING`**Provider name**: `dataDeletionJobArn`**Description**: The Amazon Resource Name (ARN) of the data deletion job.

## `data_source`{% #data_source %}

**Type**: `STRUCT`**Provider name**: `dataSource`

- `data_location`**Type**: `STRING`**Provider name**: `dataLocation`**Description**: For dataset import jobs, the path to the Amazon S3 bucket where the data that you want to upload to your dataset is stored. For data deletion jobs, the path to the Amazon S3 bucket that stores the list of records to delete. For example: `s3://bucket-name/folder-name/fileName.csv` If your CSV files are in a folder in your Amazon S3 bucket and you want your import job or data deletion job to consider multiple files, you can specify the path to the folder. With a data deletion job, Amazon Personalize uses all files in the folder and any sub folder. Use the following syntax with a `/` after the folder name: `s3://bucket-name/folder-name/`

## `dataset_group_arn`{% #dataset_group_arn %}

**Type**: `STRING`**Provider name**: `datasetGroupArn`**Description**: The Amazon Resource Name (ARN) of the dataset group the job deletes records from.

## `failure_reason`{% #failure_reason %}

**Type**: `STRING`**Provider name**: `failureReason`**Description**: If a data deletion job fails, provides the reason why.

## `job_name`{% #job_name %}

**Type**: `STRING`**Provider name**: `jobName`**Description**: The name of the data deletion job.

## `last_updated_date_time`{% #last_updated_date_time %}

**Type**: `TIMESTAMP`**Provider name**: `lastUpdatedDateTime`**Description**: The date and time (in Unix time) the data deletion job was last updated.

## `num_deleted`{% #num_deleted %}

**Type**: `INT32`**Provider name**: `numDeleted`**Description**: The number of records deleted by a COMPLETED job.

## `role_arn`{% #role_arn %}

**Type**: `STRING`**Provider name**: `roleArn`**Description**: The Amazon Resource Name (ARN) of the IAM role that has permissions to read from the Amazon S3 data source.

## `status`{% #status %}

**Type**: `STRING`**Provider name**: `status`**Description**: The status of the data deletion job. A data deletion job can have one of the following statuses:

- PENDING > IN_PROGRESS > COMPLETED -or- FAILED



## `tags`{% #tags %}

**Type**: `UNORDERED_LIST_STRING`
