This product is not supported for your selected Datadog site. ().
Cette page n'est pas encore disponible en français, sa traduction est en cours.
Si vous avez des questions ou des retours sur notre projet de traduction actuel, n'hésitez pas à nous contacter.

aws_bedrock_model_invocation_job

account_id

Type: STRING

client_request_token

Type: STRING
Provider name: clientRequestToken
Description: A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.

end_time

Type: TIMESTAMP
Provider name: endTime
Description: The time at which the batch inference job ended.

input_data_config

Type: STRUCT
Provider name: inputDataConfig
Description: Details about the location of the input to the batch inference job.

  • s3_input_data_config
    Type: STRUCT
    Provider name: s3InputDataConfig
    Description: Contains the configuration of the S3 location of the input data.
    • s3_bucket_owner
      Type: STRING
      Provider name: s3BucketOwner
      Description: The ID of the Amazon Web Services account that owns the S3 bucket containing the input data.
    • s3_input_format
      Type: STRING
      Provider name: s3InputFormat
      Description: The format of the input data.
    • s3_uri
      Type: STRING
      Provider name: s3Uri
      Description: The S3 location of the input data.

job_arn

Type: STRING
Provider name: jobArn
Description: The Amazon Resource Name (ARN) of the batch inference job.

job_expiration_time

Type: TIMESTAMP
Provider name: jobExpirationTime
Description: The time at which the batch inference job times or timed out.

job_name

Type: STRING
Provider name: jobName
Description: The name of the batch inference job.

last_modified_time

Type: TIMESTAMP
Provider name: lastModifiedTime
Description: The time at which the batch inference job was last modified.

message

Type: STRING
Provider name: message
Description: If the batch inference job failed, this field contains a message describing why the job failed.

model_id

Type: STRING
Provider name: modelId
Description: The unique identifier of the foundation model used for model inference.

output_data_config

Type: STRUCT
Provider name: outputDataConfig
Description: Details about the location of the output of the batch inference job.

  • s3_output_data_config
    Type: STRUCT
    Provider name: s3OutputDataConfig
    Description: Contains the configuration of the S3 location of the output data.
    • s3_bucket_owner
      Type: STRING
      Provider name: s3BucketOwner
      Description: The ID of the Amazon Web Services account that owns the S3 bucket containing the output data.
    • s3_encryption_key_id
      Type: STRING
      Provider name: s3EncryptionKeyId
      Description: The unique identifier of the key that encrypts the S3 location of the output data.
    • s3_uri
      Type: STRING
      Provider name: s3Uri
      Description: The S3 location of the output data.

role_arn

Type: STRING
Provider name: roleArn
Description: The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.

status

Type: STRING
Provider name: status
Description: The status of the batch inference job. The following statuses are possible:

  • Submitted – This job has been submitted to a queue for validation.
  • Validating – This job is being validated for the requirements described in Format and upload your batch inference data. The criteria include the following:
    • Your IAM service role has access to the Amazon S3 buckets containing your files.
    • Your files are .jsonl files and each individual record is a JSON object in the correct format. Note that validation doesn’t check if the modelInput value matches the request body for the model.
    • Your files fulfill the requirements for file size and number of records. For more information, see Quotas for Amazon Bedrock.
  • Scheduled – This job has been validated and is now in a queue. The job will automatically start when it reaches its turn.
  • Expired – This job timed out because it was scheduled but didn’t begin before the set timeout duration. Submit a new job request.
  • InProgress – This job has begun. You can start viewing the results in the output S3 location.
  • Completed – This job has successfully completed. View the output files in the output S3 location.
  • PartiallyCompleted – This job has partially completed. Not all of your records could be processed in time. View the output files in the output S3 location.
  • Failed – This job has failed. Check the failure message for any further details. For further assistance, reach out to the Amazon Web ServicesSupport Center.
  • Stopped – This job was stopped by a user.
  • Stopping – This job is being stopped by a user.

submit_time

Type: TIMESTAMP
Provider name: submitTime
Description: The time at which the batch inference job was submitted.

tags

Type: UNORDERED_LIST_STRING

timeout_duration_in_hours

Type: INT32
Provider name: timeoutDurationInHours
Description: The number of hours after which batch inference job was set to time out.

vpc_config

Type: STRUCT
Provider name: vpcConfig
Description: The configuration of the Virtual Private Cloud (VPC) for the data in the batch inference job. For more information, see Protect batch inference jobs using a VPC.

  • security_group_ids
    Type: UNORDERED_LIST_STRING
    Provider name: securityGroupIds
    Description: An array of IDs for each security group in the VPC to use.
  • subnet_ids
    Type: UNORDERED_LIST_STRING
    Provider name: subnetIds
    Description: An array of IDs for each subnet in the VPC to use.