Dataplex Task

A Dataplex Task in Google Cloud is a scheduled or on-demand job that runs data processing workloads within a Dataplex lake or zone. It is commonly used to automate data quality checks, transformations, or other custom operations using Spark or other supported execution environments. Tasks help maintain data consistency, enforce governance, and streamline data preparation across distributed datasets.

gcp.dataplex_task

Fields

TitleIDTypeData TypeDescription
_keycorestring
ancestorscorearray<string>
create_timecoretimestampOutput only. The time when the task was created.
datadog_display_namecorestring
descriptioncorestringOptional. Description of the task.
execution_speccorejsonRequired. Spec related to how a task is executed.
execution_statuscorejsonOutput only. Status of the latest task executions.
gcp_display_namecorestringOptional. User friendly display name.
labelscorearray<string>Optional. User-defined labels for the task.
namecorestringOutput only. The relative resource name of the task, of the form: projects/{project_number}/locations/{location_id}/lakes/{lake_id}/ tasks/{task_id}.
notebookcorejsonConfig related to running scheduled Notebooks.
organization_idcorestring
parentcorestring
project_idcorestring
project_numbercorestring
resource_namecorestring
sparkcorejsonConfig related to running custom Spark tasks.
statecorestringOutput only. Current state of the task.
tagscorehstore
trigger_speccorejsonRequired. Spec related to how often and when a task should be triggered.
uidcorestringOutput only. System generated globally unique ID for the task. This ID will be different if the task is deleted and re-created with the same name.
update_timecoretimestampOutput only. The time when the task was last updated.