Bedrock / Client / create_model_invocation_job

create_model_invocation_job#

Bedrock.Client.create_model_invocation_job(**kwargs)#

Creates a job to invoke a model on multiple prompts (batch inference). Format your data according to Format your inference data and upload it to an Amazon S3 bucket. For more information, see Create a batch inference job.

The response returns a jobArn that you can use to stop or get details about the job. You can check the status of the job by sending a GetModelCustomizationJob request.

See also: AWS API Documentation

Request Syntax

response = client.create_model_invocation_job(
    jobName='string',
    roleArn='string',
    clientRequestToken='string',
    modelId='string',
    inputDataConfig={
        's3InputDataConfig': {
            's3InputFormat': 'JSONL',
            's3Uri': 'string'
        }
    },
    outputDataConfig={
        's3OutputDataConfig': {
            's3Uri': 'string',
            's3EncryptionKeyId': 'string'
        }
    },
    timeoutDurationInHours=123,
    tags=[
        {
            'key': 'string',
            'value': 'string'
        },
    ]
)
Parameters:
  • jobName (string) –

    [REQUIRED]

    A name to give the batch inference job.

  • roleArn (string) –

    [REQUIRED]

    The Amazon Resource Name (ARN) of the service role with permissions to carry out and manage batch inference. You can use the console to create a default service role or follow the steps at Create a service role for batch inference.

  • clientRequestToken (string) –

    A unique, case-sensitive identifier to ensure that the API request completes no more than one time. If this token matches a previous request, Amazon Bedrock ignores the request, but does not return an error. For more information, see Ensuring idempotency.

    This field is autopopulated if not provided.

  • modelId (string) –

    [REQUIRED]

    The unique identifier of the foundation model to use for the batch inference job.

  • inputDataConfig (dict) –

    [REQUIRED]

    Details about the location of the input to the batch inference job.

    Note

    This is a Tagged Union structure. Only one of the following top level keys can be set: s3InputDataConfig.

    • s3InputDataConfig (dict) –

      Contains the configuration of the S3 location of the input data.

      • s3InputFormat (string) –

        The format of the input data.

      • s3Uri (string) – [REQUIRED]

        The S3 location of the input data.

  • outputDataConfig (dict) –

    [REQUIRED]

    Details about the location of the output of the batch inference job.

    Note

    This is a Tagged Union structure. Only one of the following top level keys can be set: s3OutputDataConfig.

    • s3OutputDataConfig (dict) –

      Contains the configuration of the S3 location of the output data.

      • s3Uri (string) – [REQUIRED]

        The S3 location of the output data.

      • s3EncryptionKeyId (string) –

        The unique identifier of the key that encrypts the S3 location of the output data.

  • timeoutDurationInHours (integer) – The number of hours after which to force the batch inference job to time out.

  • tags (list) –

    Any tags to associate with the batch inference job. For more information, see Tagging Amazon Bedrock resources.

    • (dict) –

      Definition of the key/value pair for a tag.

      • key (string) – [REQUIRED]

        Key for the tag.

      • value (string) – [REQUIRED]

        Value for the tag.

Return type:

dict

Returns:

Response Syntax

{
    'jobArn': 'string'
}

Response Structure

  • (dict) –

    • jobArn (string) –

      The Amazon Resource Name (ARN) of the batch inference job.

Exceptions

  • Bedrock.Client.exceptions.ResourceNotFoundException

  • Bedrock.Client.exceptions.AccessDeniedException

  • Bedrock.Client.exceptions.ValidationException

  • Bedrock.Client.exceptions.ConflictException

  • Bedrock.Client.exceptions.InternalServerException

  • Bedrock.Client.exceptions.ServiceQuotaExceededException

  • Bedrock.Client.exceptions.ThrottlingException