CleanRoomsML / Client / start_trained_model_inference_job

start_trained_model_inference_job#

CleanRoomsML.Client.start_trained_model_inference_job(**kwargs)#

Defines the information necessary to begin a trained model inference job.

See also: AWS API Documentation

Request Syntax

response = client.start_trained_model_inference_job(
    membershipIdentifier='string',
    name='string',
    trainedModelArn='string',
    configuredModelAlgorithmAssociationArn='string',
    resourceConfig={
        'instanceType': 'ml.r7i.48xlarge'|'ml.r6i.16xlarge'|'ml.m6i.xlarge'|'ml.m5.4xlarge'|'ml.p2.xlarge'|'ml.m4.16xlarge'|'ml.r7i.16xlarge'|'ml.m7i.xlarge'|'ml.m6i.12xlarge'|'ml.r7i.8xlarge'|'ml.r7i.large'|'ml.m7i.12xlarge'|'ml.m6i.24xlarge'|'ml.m7i.24xlarge'|'ml.r6i.8xlarge'|'ml.r6i.large'|'ml.g5.2xlarge'|'ml.m5.large'|'ml.p3.16xlarge'|'ml.m7i.48xlarge'|'ml.m6i.16xlarge'|'ml.p2.16xlarge'|'ml.g5.4xlarge'|'ml.m7i.16xlarge'|'ml.c4.2xlarge'|'ml.c5.2xlarge'|'ml.c6i.32xlarge'|'ml.c4.4xlarge'|'ml.g5.8xlarge'|'ml.c6i.xlarge'|'ml.c5.4xlarge'|'ml.g4dn.xlarge'|'ml.c7i.xlarge'|'ml.c6i.12xlarge'|'ml.g4dn.12xlarge'|'ml.c7i.12xlarge'|'ml.c6i.24xlarge'|'ml.g4dn.2xlarge'|'ml.c7i.24xlarge'|'ml.c7i.2xlarge'|'ml.c4.8xlarge'|'ml.c6i.2xlarge'|'ml.g4dn.4xlarge'|'ml.c7i.48xlarge'|'ml.c7i.4xlarge'|'ml.c6i.16xlarge'|'ml.c5.9xlarge'|'ml.g4dn.16xlarge'|'ml.c7i.16xlarge'|'ml.c6i.4xlarge'|'ml.c5.xlarge'|'ml.c4.xlarge'|'ml.g4dn.8xlarge'|'ml.c7i.8xlarge'|'ml.c7i.large'|'ml.g5.xlarge'|'ml.c6i.8xlarge'|'ml.c6i.large'|'ml.g5.12xlarge'|'ml.g5.24xlarge'|'ml.m7i.2xlarge'|'ml.c5.18xlarge'|'ml.g5.48xlarge'|'ml.m6i.2xlarge'|'ml.g5.16xlarge'|'ml.m7i.4xlarge'|'ml.p3.2xlarge'|'ml.r6i.32xlarge'|'ml.m6i.4xlarge'|'ml.m5.xlarge'|'ml.m4.10xlarge'|'ml.r6i.xlarge'|'ml.m5.12xlarge'|'ml.m4.xlarge'|'ml.r7i.2xlarge'|'ml.r7i.xlarge'|'ml.r6i.12xlarge'|'ml.m5.24xlarge'|'ml.r7i.12xlarge'|'ml.m7i.8xlarge'|'ml.m7i.large'|'ml.r6i.24xlarge'|'ml.r6i.2xlarge'|'ml.m4.2xlarge'|'ml.r7i.24xlarge'|'ml.r7i.4xlarge'|'ml.m6i.8xlarge'|'ml.m6i.large'|'ml.m5.2xlarge'|'ml.p2.8xlarge'|'ml.r6i.4xlarge'|'ml.m6i.32xlarge'|'ml.p3.8xlarge'|'ml.m4.4xlarge',
        'instanceCount': 123
    },
    outputConfiguration={
        'accept': 'string',
        'members': [
            {
                'accountId': 'string'
            },
        ]
    },
    dataSource={
        'mlInputChannelArn': 'string'
    },
    description='string',
    containerExecutionParameters={
        'maxPayloadInMB': 123
    },
    environment={
        'string': 'string'
    },
    kmsKeyArn='string',
    tags={
        'string': 'string'
    }
)
Parameters:
  • membershipIdentifier (string) –

    [REQUIRED]

    The membership ID of the membership that contains the trained model inference job.

  • name (string) –

    [REQUIRED]

    The name of the trained model inference job.

  • trainedModelArn (string) –

    [REQUIRED]

    The Amazon Resource Name (ARN) of the trained model that is used for this trained model inference job.

  • configuredModelAlgorithmAssociationArn (string) – The Amazon Resource Name (ARN) of the configured model algorithm association that is used for this trained model inference job.

  • resourceConfig (dict) –

    [REQUIRED]

    Defines the resource configuration for the trained model inference job.

    • instanceType (string) – [REQUIRED]

      The type of instance that is used to perform model inference.

    • instanceCount (integer) –

      The number of instances to use.

  • outputConfiguration (dict) –

    [REQUIRED]

    Defines the output configuration information for the trained model inference job.

    • accept (string) –

      The MIME type used to specify the output data.

    • members (list) – [REQUIRED]

      Defines the members that can receive inference output.

      • (dict) –

        Defines who will receive inference results.

        • accountId (string) – [REQUIRED]

          The account ID of the member that can receive inference results.

  • dataSource (dict) –

    [REQUIRED]

    Defines the data source that is used for the trained model inference job.

    • mlInputChannelArn (string) – [REQUIRED]

      The Amazon Resource Name (ARN) of the ML input channel for this model inference data source.

  • description (string) – The description of the trained model inference job.

  • containerExecutionParameters (dict) –

    The execution parameters for the container.

    • maxPayloadInMB (integer) –

      The maximum size of the inference container payload, specified in MB.

  • environment (dict) –

    The environment variables to set in the Docker container.

    • (string) –

      • (string) –

  • kmsKeyArn (string) – The Amazon Resource Name (ARN) of the KMS key. This key is used to encrypt and decrypt customer-owned data in the ML inference job and associated data.

  • tags (dict) –

    The optional metadata that you apply to the resource to help you categorize and organize them. Each tag consists of a key and an optional value, both of which you define.

    The following basic restrictions apply to tags:

    • Maximum number of tags per resource - 50.

    • For each resource, each tag key must be unique, and each tag key can have only one value.

    • Maximum key length - 128 Unicode characters in UTF-8.

    • Maximum value length - 256 Unicode characters in UTF-8.

    • If your tagging schema is used across multiple services and resources, remember that other services may have restrictions on allowed characters. Generally allowed characters are: letters, numbers, and spaces representable in UTF-8, and the following characters: + - = . _ : / @.

    • Tag keys and values are case sensitive.

    • Do not use aws:, AWS:, or any upper or lowercase combination of such as a prefix for keys as it is reserved for AWS use. You cannot edit or delete tag keys with this prefix. Values can have this prefix. If a tag value has aws as its prefix but the key does not, then Clean Rooms ML considers it to be a user tag and will count against the limit of 50 tags. Tags with only the key prefix of aws do not count against your tags per resource limit.

    • (string) –

      • (string) –

Return type:

dict

Returns:

Response Syntax

{
    'trainedModelInferenceJobArn': 'string'
}

Response Structure

  • (dict) –

    • trainedModelInferenceJobArn (string) –

      The Amazon Resource Name (ARN) of the trained model inference job.

Exceptions

  • CleanRoomsML.Client.exceptions.ConflictException

  • CleanRoomsML.Client.exceptions.ValidationException

  • CleanRoomsML.Client.exceptions.AccessDeniedException

  • CleanRoomsML.Client.exceptions.ResourceNotFoundException

  • CleanRoomsML.Client.exceptions.ServiceQuotaExceededException