- 필수 기능
- 시작하기
- Glossary
- 표준 속성
- Guides
- Agent
- 통합
- 개방형텔레메트리
- 개발자
- Administrator's Guide
- API
- Datadog Mobile App
- CoScreen
- Cloudcraft
- 앱 내
- 서비스 관리
- 인프라스트럭처
- 애플리케이션 성능
- APM
- Continuous Profiler
- 스팬 시각화
- 데이터 스트림 모니터링
- 데이터 작업 모니터링
- 디지털 경험
- 소프트웨어 제공
- 보안
- AI Observability
- 로그 관리
- 관리
",t};e.buildCustomizationMenuUi=t;function n(e){let t='
",t}function s(e){let n=e.filter.currentValue||e.filter.defaultValue,t='${e.filter.label}
`,e.filter.options.forEach(s=>{let o=s.id===n;t+=``}),t+="${e.filter.label}
`,t+=`ancestors
Type: UNORDERED_LIST_STRING
client_request_id
Type: STRING
Provider name: clientRequestId
Description: The client’s unique identifier of the job, re-used across retried attempts. If this field is set, the service will ensure its uniqueness. The request to create a job will fail if the service has knowledge of a previously submitted job with the same client’s ID and job name. The caller may use this field to ensure idempotence of job creation across retried attempts to create a job. By default, the field is empty and, in that case, the service ignores it.
create_time
Type: TIMESTAMP
Provider name: createTime
Description: The timestamp when the job was initially created. Immutable and set by the Cloud Dataflow service.
created_from_snapshot_id
Type: STRING
Provider name: createdFromSnapshotId
Description: If this is specified, the job’s initial state is populated from the given snapshot.
current_state
Type: STRING
Provider name: currentState
Description: The current state of the job. Jobs are created in the JOB_STATE_STOPPED
state unless otherwise specified. A job in the JOB_STATE_RUNNING
state may asynchronously enter a terminal state. After a job has reached a terminal state, no further state updates may be made. This field might be mutated by the Dataflow service; callers cannot mutate it.
Possible values:
JOB_STATE_UNKNOWN
- The job’s run state isn’t specified.JOB_STATE_STOPPED
- JOB_STATE_STOPPED
indicates that the job has not yet started to run.JOB_STATE_RUNNING
- JOB_STATE_RUNNING
indicates that the job is currently running.JOB_STATE_DONE
- JOB_STATE_DONE
indicates that the job has successfully completed. This is a terminal job state. This state may be set by the Cloud Dataflow service, as a transition from JOB_STATE_RUNNING
. It may also be set via a Cloud Dataflow UpdateJob
call, if the job has not yet reached a terminal state.JOB_STATE_FAILED
- JOB_STATE_FAILED
indicates that the job has failed. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_CANCELLED
- JOB_STATE_CANCELLED
indicates that the job has been explicitly cancelled. This is a terminal job state. This state may only be set via a Cloud Dataflow UpdateJob
call, and only if the job has not yet reached another terminal state.JOB_STATE_UPDATED
- JOB_STATE_UPDATED
indicates that the job was successfully updated, meaning that this job was stopped and another job was started, inheriting state from this one. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_DRAINING
- JOB_STATE_DRAINING
indicates that the job is in the process of draining. A draining job has stopped pulling from its input sources and is processing any data that remains in-flight. This state may be set via a Cloud Dataflow UpdateJob
call, but only as a transition from JOB_STATE_RUNNING
. Jobs that are draining may only transition to JOB_STATE_DRAINED
, JOB_STATE_CANCELLED
, or JOB_STATE_FAILED
.JOB_STATE_DRAINED
- JOB_STATE_DRAINED
indicates that the job has been drained. A drained job terminated by stopping pulling from its input sources and processing any data that remained in-flight when draining was requested. This state is a terminal state, may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_DRAINING
.JOB_STATE_PENDING
- JOB_STATE_PENDING
indicates that the job has been created but is not yet running. Jobs that are pending may only transition to JOB_STATE_RUNNING
, or JOB_STATE_FAILED
.JOB_STATE_CANCELLING
- JOB_STATE_CANCELLING
indicates that the job has been explicitly cancelled and is in the process of stopping. Jobs that are cancelling may only transition to JOB_STATE_CANCELLED
or JOB_STATE_FAILED
.JOB_STATE_QUEUED
- JOB_STATE_QUEUED
indicates that the job has been created but is being delayed until launch. Jobs that are queued may only transition to JOB_STATE_PENDING
or JOB_STATE_CANCELLED
.JOB_STATE_RESOURCE_CLEANING_UP
- JOB_STATE_RESOURCE_CLEANING_UP
indicates that the batch job’s associated resources are currently being cleaned up after a successful run. Currently, this is an opt-in feature, please reach out to Cloud support team if you are interested.current_state_time
Type: TIMESTAMP
Provider name: currentStateTime
Description: The timestamp associated with the current state.
environment
Type: STRUCT
Provider name: environment
Description: Optional. The environment for the job.
cluster_manager_api_service
STRING
clusterManagerApiService
dataset
STRING
dataset
debug_options
STRUCT
debugOptions
data_sampling
STRUCT
dataSampling
behaviors
UNORDERED_LIST_STRING
behaviors
enable_hot_key_logging
BOOLEAN
enableHotKeyLogging
experiments
UNORDERED_LIST_STRING
experiments
flex_resource_scheduling_goal
STRING
flexResourceSchedulingGoal
FLEXRS_UNSPECIFIED
- Run in the default mode.FLEXRS_SPEED_OPTIMIZED
- Optimize for lower execution time.FLEXRS_COST_OPTIMIZED
- Optimize for lower cost.service_account_email
STRING
serviceAccountEmail
service_kms_key_name
STRING
serviceKmsKeyName
service_options
UNORDERED_LIST_STRING
serviceOptions
shuffle_mode
STRING
shuffleMode
SHUFFLE_MODE_UNSPECIFIED
- Shuffle mode information is not available.VM_BASED
- Shuffle is done on the worker VMs.SERVICE_BASED
- Shuffle is done on the service side.streaming_mode
STRING
streamingMode
STREAMING_MODE_UNSPECIFIED
- Run in the default mode.STREAMING_MODE_EXACTLY_ONCE
- In this mode, message deduplication is performed against persistent state to make sure each message is processed and committed to storage exactly once.STREAMING_MODE_AT_LEAST_ONCE
- Message deduplication is not performed. Messages might be processed multiple times, and the results are applied multiple times. Note: Setting this value also enables Streaming Engine and Streaming Engine resource-based billing.temp_storage_prefix
STRING
tempStoragePrefix
use_streaming_engine_resource_based_billing
BOOLEAN
useStreamingEngineResourceBasedBilling
worker_pools
UNORDERED_LIST_STRUCT
workerPools
autoscaling_settings
STRUCT
autoscalingSettings
algorithm
STRING
algorithm
AUTOSCALING_ALGORITHM_UNKNOWN
- The algorithm is unknown, or unspecified.AUTOSCALING_ALGORITHM_NONE
- Disable autoscaling.AUTOSCALING_ALGORITHM_BASIC
- Increase worker count over time to reduce job execution time.max_num_workers
INT32
maxNumWorkers
data_disks
UNORDERED_LIST_STRUCT
dataDisks
disk_type
STRING
diskType
mount_point
STRING
mountPoint
size_gb
INT32
sizeGb
default_package_set
STRING
defaultPackageSet
DEFAULT_PACKAGE_SET_UNKNOWN
- The default set of packages to stage is unknown, or unspecified.DEFAULT_PACKAGE_SET_NONE
- Indicates that no packages should be staged at the worker unless explicitly specified by the job.DEFAULT_PACKAGE_SET_JAVA
- Stage packages typically useful to workers written in Java.DEFAULT_PACKAGE_SET_PYTHON
- Stage packages typically useful to workers written in Python.disk_size_gb
INT32
diskSizeGb
disk_source_image
STRING
diskSourceImage
disk_type
STRING
diskType
ip_configuration
STRING
ipConfiguration
WORKER_IP_UNSPECIFIED
- The configuration is unknown, or unspecified.WORKER_IP_PUBLIC
- Workers should have public IP addresses.WORKER_IP_PRIVATE
- Workers should have private IP addresses.kind
STRING
kind
harness
and shuffle
are supported.machine_type
STRING
machineType
network
STRING
network
num_threads_per_worker
INT32
numThreadsPerWorker
num_workers
INT32
numWorkers
on_host_maintenance
STRING
onHostMaintenance
packages
UNORDERED_LIST_STRUCT
packages
location
STRING
location
name
STRING
name
sdk_harness_container_images
UNORDERED_LIST_STRUCT
sdkHarnessContainerImages
capabilities
UNORDERED_LIST_STRING
capabilities
container_image
STRING
containerImage
environment_id
STRING
environmentId
use_single_core_per_container
BOOLEAN
useSingleCorePerContainer
subnetwork
STRING
subnetwork
taskrunner_settings
STRUCT
taskrunnerSettings
alsologtostderr
BOOLEAN
alsologtostderr
base_task_dir
STRING
baseTaskDir
base_url
STRING
baseUrl
commandlines_file_name
STRING
commandlinesFileName
continue_on_exception
BOOLEAN
continueOnException
dataflow_api_version
STRING
dataflowApiVersion
harness_command
STRING
harnessCommand
language_hint
STRING
languageHint
log_dir
STRING
logDir
log_to_serialconsole
BOOLEAN
logToSerialconsole
log_upload_location
STRING
logUploadLocation
oauth_scopes
UNORDERED_LIST_STRING
oauthScopes
parallel_worker_settings
STRUCT
parallelWorkerSettings
base_url
STRING
baseUrl
reporting_enabled
BOOLEAN
reportingEnabled
service_path
STRING
servicePath
shuffle_service_path
STRING
shuffleServicePath
temp_storage_prefix
STRING
tempStoragePrefix
worker_id
STRING
workerId
streaming_worker_main_class
STRING
streamingWorkerMainClass
task_group
STRING
taskGroup
task_user
STRING
taskUser
temp_storage_prefix
STRING
tempStoragePrefix
vm_id
STRING
vmId
workflow_file_name
STRING
workflowFileName
teardown_policy
STRING
teardownPolicy
TEARDOWN_ALWAYS
, TEARDOWN_ON_SUCCESS
, and TEARDOWN_NEVER
. TEARDOWN_ALWAYS
means workers are always torn down regardless of whether the job succeeds. TEARDOWN_ON_SUCCESS
means workers are torn down if the job succeeds. TEARDOWN_NEVER
means the workers are never torn down. If the workers are not torn down by the service, they will continue to run and use Google Compute Engine VM resources in the user’s project until they are explicitly terminated by the user. Because of this, Google recommends using the TEARDOWN_ALWAYS
policy except for small, manually supervised test jobs. If unknown or unspecified, the service will attempt to choose a reasonable default.TEARDOWN_POLICY_UNKNOWN
- The teardown policy isn’t specified, or is unknown.TEARDOWN_ALWAYS
- Always teardown the resource.TEARDOWN_ON_SUCCESS
- Teardown the resource on success. This is useful for debugging failures.TEARDOWN_NEVER
- Never teardown the resource. This is useful for debugging and development.worker_harness_container_image
STRING
workerHarnessContainerImage
zone
STRING
zone
worker_region
STRING
workerRegion
worker_zone
STRING
workerZone
execution_info
Type: STRUCT
Provider name: executionInfo
Description: Deprecated.
id
Type: STRING
Provider name: id
Description: The unique ID of this job. This field is set by the Dataflow service when the job is created, and is immutable for the life of the job.
job_metadata
Type: STRUCT
Provider name: jobMetadata
Description: This field is populated by the Dataflow service to support filtering jobs by the metadata values provided here. Populated for ListJobs and all GetJob views SUMMARY and higher.
big_table_details
UNORDERED_LIST_STRUCT
bigTableDetails
instance_id
STRING
instanceId
project_id
STRING
projectId
table_id
STRING
tableId
bigquery_details
UNORDERED_LIST_STRUCT
bigqueryDetails
dataset
STRING
dataset
project_id
STRING
projectId
query
STRING
query
table
STRING
table
datastore_details
UNORDERED_LIST_STRUCT
datastoreDetails
namespace
STRING
namespace
project_id
STRING
projectId
file_details
UNORDERED_LIST_STRUCT
fileDetails
file_pattern
STRING
filePattern
pubsub_details
UNORDERED_LIST_STRUCT
pubsubDetails
subscription
STRING
subscription
topic
STRING
topic
sdk_version
STRUCT
sdkVersion
bugs
UNORDERED_LIST_STRUCT
bugs
severity
STRING
severity
SEVERITY_UNSPECIFIED
- A bug of unknown severity.NOTICE
- A minor bug that that may reduce reliability or performance for some jobs. Impact will be minimal or non-existent for most jobs.WARNING
- A bug that has some likelihood of causing performance degradation, data loss, or job failures.SEVERE
- A bug with extremely significant impact. Jobs may fail erroneously, performance may be severely degraded, and data loss may be very likely.type
STRING
type
TYPE_UNSPECIFIED
- Unknown issue with this SDK.GENERAL
- Catch-all for SDK bugs that don’t fit in the below categories.PERFORMANCE
- Using this version of the SDK may result in degraded performance.DATALOSS
- Using this version of the SDK may cause data loss.uri
STRING
uri
sdk_support_status
STRING
sdkSupportStatus
UNKNOWN
- Cloud Dataflow is unaware of this version.SUPPORTED
- This is a known version of an SDK, and is supported.STALE
- A newer version of the SDK family exists, and an update is recommended.DEPRECATED
- This version of the SDK is deprecated and will eventually be unsupported.UNSUPPORTED
- Support for this SDK version has ended and it should no longer be used.version
STRING
version
version_display_name
STRING
versionDisplayName
spanner_details
UNORDERED_LIST_STRUCT
spannerDetails
database_id
STRING
databaseId
instance_id
STRING
instanceId
project_id
STRING
projectId
labels
Type: UNORDERED_LIST_STRING
location
Type: STRING
Provider name: location
Description: Optional. The [regional endpoint] (https://cloud.google.com/dataflow/docs/concepts/regional-endpoints) that contains this job.
name
Type: STRING
Provider name: name
Description: Optional. The user-specified Dataflow job name. Only one active job with a given name can exist in a project within one region at any given time. Jobs in different regions can have the same name. If a caller attempts to create a job with the same name as an active job that already exists, the attempt returns the existing job. The name must match the regular expression [a-z]([-a-z0-9]{0,1022}[a-z0-9])?
organization_id
Type: STRING
parent
Type: STRING
pipeline_description
Type: STRUCT
Provider name: pipelineDescription
Description: Preliminary field: The format of this data may change at any time. A description of the user pipeline and stages through which it is executed. Created by Cloud Dataflow service. Only retrieved with JOB_VIEW_DESCRIPTION or JOB_VIEW_ALL.
display_data
UNORDERED_LIST_STRUCT
displayData
bool_value
BOOLEAN
boolValue
duration_value
STRING
durationValue
float_value
FLOAT
floatValue
int64_value
INT64
int64Value
java_class_value
STRING
javaClassValue
key
STRING
key
label
STRING
label
namespace
STRING
namespace
short_str_value
STRING
shortStrValue
str_value
STRING
strValue
timestamp_value
TIMESTAMP
timestampValue
url
STRING
url
execution_pipeline_stage
UNORDERED_LIST_STRUCT
executionPipelineStage
component_source
UNORDERED_LIST_STRUCT
componentSource
name
STRING
name
original_transform_or_collection
STRING
originalTransformOrCollection
user_name
STRING
userName
component_transform
UNORDERED_LIST_STRUCT
componentTransform
name
STRING
name
original_transform
STRING
originalTransform
user_name
STRING
userName
id
STRING
id
input_source
UNORDERED_LIST_STRUCT
inputSource
name
STRING
name
original_transform_or_collection
STRING
originalTransformOrCollection
size_bytes
INT64
sizeBytes
user_name
STRING
userName
kind
STRING
kind
UNKNOWN_KIND
- Unrecognized transform type.PAR_DO_KIND
- ParDo transform.GROUP_BY_KEY_KIND
- Group By Key transform.FLATTEN_KIND
- Flatten transform.READ_KIND
- Read transform.WRITE_KIND
- Write transform.CONSTANT_KIND
- Constructs from a constant value, such as with Create.of.SINGLETON_KIND
- Creates a Singleton view of a collection.SHUFFLE_KIND
- Opening or closing a shuffle session, often as part of a GroupByKey.name
STRING
name
output_source
UNORDERED_LIST_STRUCT
outputSource
name
STRING
name
original_transform_or_collection
STRING
originalTransformOrCollection
size_bytes
INT64
sizeBytes
user_name
STRING
userName
prerequisite_stage
UNORDERED_LIST_STRING
prerequisiteStage
original_pipeline_transform
UNORDERED_LIST_STRUCT
originalPipelineTransform
display_data
UNORDERED_LIST_STRUCT
displayData
bool_value
BOOLEAN
boolValue
duration_value
STRING
durationValue
float_value
FLOAT
floatValue
int64_value
INT64
int64Value
java_class_value
STRING
javaClassValue
key
STRING
key
label
STRING
label
namespace
STRING
namespace
short_str_value
STRING
shortStrValue
str_value
STRING
strValue
timestamp_value
TIMESTAMP
timestampValue
url
STRING
url
id
STRING
id
input_collection_name
UNORDERED_LIST_STRING
inputCollectionName
kind
STRING
kind
UNKNOWN_KIND
- Unrecognized transform type.PAR_DO_KIND
- ParDo transform.GROUP_BY_KEY_KIND
- Group By Key transform.FLATTEN_KIND
- Flatten transform.READ_KIND
- Read transform.WRITE_KIND
- Write transform.CONSTANT_KIND
- Constructs from a constant value, such as with Create.of.SINGLETON_KIND
- Creates a Singleton view of a collection.SHUFFLE_KIND
- Opening or closing a shuffle session, often as part of a GroupByKey.name
STRING
name
output_collection_name
UNORDERED_LIST_STRING
outputCollectionName
step_names_hash
STRING
stepNamesHash
project_id
Type: STRING
project_number
Type: STRING
replace_job_id
Type: STRING
Provider name: replaceJobId
Description: If this job is an update of an existing job, this field is the job ID of the job it replaced. When sending a CreateJobRequest
, you can update a job by specifying it here. The job named here is stopped, and its intermediate state is transferred to this job.
replaced_by_job_id
Type: STRING
Provider name: replacedByJobId
Description: If another job is an update of this job (and thus, this job is in JOB_STATE_UPDATED
), this field contains the ID of that job.
requested_state
Type: STRING
Provider name: requestedState
Description: The job’s requested state. Applies to UpdateJob
requests. Set requested_state
with UpdateJob
requests to switch between the states JOB_STATE_STOPPED
and JOB_STATE_RUNNING
. You can also use UpdateJob
requests to change a job’s state from JOB_STATE_RUNNING
to JOB_STATE_CANCELLED
, JOB_STATE_DONE
, or JOB_STATE_DRAINED
. These states irrevocably terminate the job if it hasn’t already reached a terminal state. This field has no effect on CreateJob
requests.
Possible values:
JOB_STATE_UNKNOWN
- The job’s run state isn’t specified.JOB_STATE_STOPPED
- JOB_STATE_STOPPED
indicates that the job has not yet started to run.JOB_STATE_RUNNING
- JOB_STATE_RUNNING
indicates that the job is currently running.JOB_STATE_DONE
- JOB_STATE_DONE
indicates that the job has successfully completed. This is a terminal job state. This state may be set by the Cloud Dataflow service, as a transition from JOB_STATE_RUNNING
. It may also be set via a Cloud Dataflow UpdateJob
call, if the job has not yet reached a terminal state.JOB_STATE_FAILED
- JOB_STATE_FAILED
indicates that the job has failed. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_CANCELLED
- JOB_STATE_CANCELLED
indicates that the job has been explicitly cancelled. This is a terminal job state. This state may only be set via a Cloud Dataflow UpdateJob
call, and only if the job has not yet reached another terminal state.JOB_STATE_UPDATED
- JOB_STATE_UPDATED
indicates that the job was successfully updated, meaning that this job was stopped and another job was started, inheriting state from this one. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_DRAINING
- JOB_STATE_DRAINING
indicates that the job is in the process of draining. A draining job has stopped pulling from its input sources and is processing any data that remains in-flight. This state may be set via a Cloud Dataflow UpdateJob
call, but only as a transition from JOB_STATE_RUNNING
. Jobs that are draining may only transition to JOB_STATE_DRAINED
, JOB_STATE_CANCELLED
, or JOB_STATE_FAILED
.JOB_STATE_DRAINED
- JOB_STATE_DRAINED
indicates that the job has been drained. A drained job terminated by stopping pulling from its input sources and processing any data that remained in-flight when draining was requested. This state is a terminal state, may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_DRAINING
.JOB_STATE_PENDING
- JOB_STATE_PENDING
indicates that the job has been created but is not yet running. Jobs that are pending may only transition to JOB_STATE_RUNNING
, or JOB_STATE_FAILED
.JOB_STATE_CANCELLING
- JOB_STATE_CANCELLING
indicates that the job has been explicitly cancelled and is in the process of stopping. Jobs that are cancelling may only transition to JOB_STATE_CANCELLED
or JOB_STATE_FAILED
.JOB_STATE_QUEUED
- JOB_STATE_QUEUED
indicates that the job has been created but is being delayed until launch. Jobs that are queued may only transition to JOB_STATE_PENDING
or JOB_STATE_CANCELLED
.JOB_STATE_RESOURCE_CLEANING_UP
- JOB_STATE_RESOURCE_CLEANING_UP
indicates that the batch job’s associated resources are currently being cleaned up after a successful run. Currently, this is an opt-in feature, please reach out to Cloud support team if you are interested.resource_name
Type: STRING
runtime_updatable_params
Type: STRUCT
Provider name: runtimeUpdatableParams
Description: This field may ONLY be modified at runtime using the projects.jobs.update method to adjust job behavior. This field has no effect when specified at job creation.
max_num_workers
INT32
maxNumWorkers
min_num_workers
INT32
minNumWorkers
worker_utilization_hint
DOUBLE
workerUtilizationHint
satisfies_pzi
Type: BOOLEAN
Provider name: satisfiesPzi
Description: Output only. Reserved for future use. This field is set only in responses from the server; it is ignored if it is set in any requests.
satisfies_pzs
Type: BOOLEAN
Provider name: satisfiesPzs
Description: Reserved for future use. This field is set only in responses from the server; it is ignored if it is set in any requests.
service_resources
Type: STRUCT
Provider name: serviceResources
Description: Output only. Resources used by the Dataflow Service to run the job.
zones
UNORDERED_LIST_STRING
zones
stage_states
Type: UNORDERED_LIST_STRUCT
Provider name: stageStates
Description: This field may be mutated by the Cloud Dataflow service; callers cannot mutate it.
current_state_time
TIMESTAMP
currentStateTime
execution_stage_name
STRING
executionStageName
execution_stage_state
STRING
executionStageState
JOB_STATE_UNKNOWN
- The job’s run state isn’t specified.JOB_STATE_STOPPED
- JOB_STATE_STOPPED
indicates that the job has not yet started to run.JOB_STATE_RUNNING
- JOB_STATE_RUNNING
indicates that the job is currently running.JOB_STATE_DONE
- JOB_STATE_DONE
indicates that the job has successfully completed. This is a terminal job state. This state may be set by the Cloud Dataflow service, as a transition from JOB_STATE_RUNNING
. It may also be set via a Cloud Dataflow UpdateJob
call, if the job has not yet reached a terminal state.JOB_STATE_FAILED
- JOB_STATE_FAILED
indicates that the job has failed. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_CANCELLED
- JOB_STATE_CANCELLED
indicates that the job has been explicitly cancelled. This is a terminal job state. This state may only be set via a Cloud Dataflow UpdateJob
call, and only if the job has not yet reached another terminal state.JOB_STATE_UPDATED
- JOB_STATE_UPDATED
indicates that the job was successfully updated, meaning that this job was stopped and another job was started, inheriting state from this one. This is a terminal job state. This state may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_RUNNING
.JOB_STATE_DRAINING
- JOB_STATE_DRAINING
indicates that the job is in the process of draining. A draining job has stopped pulling from its input sources and is processing any data that remains in-flight. This state may be set via a Cloud Dataflow UpdateJob
call, but only as a transition from JOB_STATE_RUNNING
. Jobs that are draining may only transition to JOB_STATE_DRAINED
, JOB_STATE_CANCELLED
, or JOB_STATE_FAILED
.JOB_STATE_DRAINED
- JOB_STATE_DRAINED
indicates that the job has been drained. A drained job terminated by stopping pulling from its input sources and processing any data that remained in-flight when draining was requested. This state is a terminal state, may only be set by the Cloud Dataflow service, and only as a transition from JOB_STATE_DRAINING
.JOB_STATE_PENDING
- JOB_STATE_PENDING
indicates that the job has been created but is not yet running. Jobs that are pending may only transition to JOB_STATE_RUNNING
, or JOB_STATE_FAILED
.JOB_STATE_CANCELLING
- JOB_STATE_CANCELLING
indicates that the job has been explicitly cancelled and is in the process of stopping. Jobs that are cancelling may only transition to JOB_STATE_CANCELLED
or JOB_STATE_FAILED
.JOB_STATE_QUEUED
- JOB_STATE_QUEUED
indicates that the job has been created but is being delayed until launch. Jobs that are queued may only transition to JOB_STATE_PENDING
or JOB_STATE_CANCELLED
.JOB_STATE_RESOURCE_CLEANING_UP
- JOB_STATE_RESOURCE_CLEANING_UP
indicates that the batch job’s associated resources are currently being cleaned up after a successful run. Currently, this is an opt-in feature, please reach out to Cloud support team if you are interested.start_time
Type: TIMESTAMP
Provider name: startTime
Description: The timestamp when the job was started (transitioned to JOB_STATE_PENDING). Flexible resource scheduling jobs are started with some delay after job creation, so start_time is unset before start and is updated when the job is started by the Cloud Dataflow service. For other jobs, start_time always equals to create_time and is immutable and set by the Cloud Dataflow service.
steps
Type: UNORDERED_LIST_STRUCT
Provider name: steps
Description: Exactly one of step or steps_location should be specified. The top-level steps that constitute the entire job. Only retrieved with JOB_VIEW_ALL.
kind
STRING
kind
name
STRING
name
steps_location
Type: STRING
Provider name: stepsLocation
Description: The Cloud Storage location where the steps are stored.
tags
Type: UNORDERED_LIST_STRING
temp_files
Type: UNORDERED_LIST_STRING
Provider name: tempFiles
Description: A set of files the system should be aware of that are used for temporary storage. These temporary files will be removed on job completion. No duplicates are allowed. No file patterns are supported. The supported files are: Google Cloud Storage: storage.googleapis.com/{bucket}/{object} bucket.storage.googleapis.com/{object}
type
Type: STRING
Provider name: type
Description: Optional. The type of Dataflow job.
Possible values:
JOB_TYPE_UNKNOWN
- The type of the job is unspecified, or unknown.JOB_TYPE_BATCH
- A batch job with a well-defined end point: data is read, data is processed, data is written, and the job is done.JOB_TYPE_STREAMING
- A continuously streaming job with no end: data is read, processed, and written continuously.