start_ml_evaluation_task_run

start_ml_evaluation_task_run(**kwargs)

Starts a task to estimate the quality of the transform.

When you provide label sets as examples of truth, Glue machine learning uses some of those examples to learn from them. The rest of the labels are used as a test to estimate quality.

Returns a unique identifier for the run. You can call GetMLTaskRun to get more information about the stats of the EvaluationTaskRun .

See also: AWS API Documentation

Request Syntax

response = client.start_ml_evaluation_task_run(
    TransformId='string'
)
Parameters
TransformId (string) --

[REQUIRED]

The unique identifier of the machine learning transform.

Return type
dict
Returns
Response Syntax
{
    'TaskRunId': 'string'
}

Response Structure

  • (dict) --
    • TaskRunId (string) --

      The unique identifier associated with this run.

Exceptions

  • Glue.Client.exceptions.EntityNotFoundException
  • Glue.Client.exceptions.InvalidInputException
  • Glue.Client.exceptions.OperationTimeoutException
  • Glue.Client.exceptions.InternalServiceException
  • Glue.Client.exceptions.ConcurrentRunsExceededException
  • Glue.Client.exceptions.MLTransformNotReadyException