---
title: Getting Started with Datadog
description: Datadog, the leading service for cloud-scale monitoring.
breadcrumbs: Docs > Infrastructure > Datadog Resource Catalog
---

# aws_personalize_dataset_export_job{% #aws_personalize_dataset_export_job %}

## `account_id`{% #account_id %}

**Type**: `STRING`

## `creation_date_time`{% #creation_date_time %}

**Type**: `TIMESTAMP`**Provider name**: `creationDateTime`**Description**: The creation date and time (in Unix time) of the dataset export job.

## `dataset_arn`{% #dataset_arn %}

**Type**: `STRING`**Provider name**: `datasetArn`**Description**: The Amazon Resource Name (ARN) of the dataset to export.

## `dataset_export_job_arn`{% #dataset_export_job_arn %}

**Type**: `STRING`**Provider name**: `datasetExportJobArn`**Description**: The Amazon Resource Name (ARN) of the dataset export job.

## `failure_reason`{% #failure_reason %}

**Type**: `STRING`**Provider name**: `failureReason`**Description**: If a dataset export job fails, provides the reason why.

## `ingestion_mode`{% #ingestion_mode %}

**Type**: `STRING`**Provider name**: `ingestionMode`**Description**: The data to export, based on how you imported the data. You can choose to export `BULK` data that you imported using a dataset import job, `PUT` data that you imported incrementally (using the console, PutEvents, PutUsers and PutItems operations), or `ALL` for both types. The default value is `PUT`.

## `job_name`{% #job_name %}

**Type**: `STRING`**Provider name**: `jobName`**Description**: The name of the export job.

## `job_output`{% #job_output %}

**Type**: `STRUCT`**Provider name**: `jobOutput`**Description**: The path to the Amazon S3 bucket where the job's output is stored. For example: `s3://bucket-name/folder-name/`

- `s3_data_destination`**Type**: `STRUCT`**Provider name**: `s3DataDestination`
  - `kms_key_arn`**Type**: `STRING`**Provider name**: `kmsKeyArn`**Description**: The Amazon Resource Name (ARN) of the Key Management Service (KMS) key that Amazon Personalize uses to encrypt or decrypt the input and output files.
  - `path`**Type**: `STRING`**Provider name**: `path`**Description**: The file path of the Amazon S3 bucket.

## `last_updated_date_time`{% #last_updated_date_time %}

**Type**: `TIMESTAMP`**Provider name**: `lastUpdatedDateTime`**Description**: The date and time (in Unix time) the status of the dataset export job was last updated.

## `role_arn`{% #role_arn %}

**Type**: `STRING`**Provider name**: `roleArn`**Description**: The Amazon Resource Name (ARN) of the IAM service role that has permissions to add data to your output Amazon S3 bucket.

## `status`{% #status %}

**Type**: `STRING`**Provider name**: `status`**Description**: The status of the dataset export job. A dataset export job can be in one of the following states:

- CREATE PENDING > CREATE IN_PROGRESS > ACTIVE -or- CREATE FAILED



## `tags`{% #tags %}

**Type**: `UNORDERED_LIST_STRING`
