Glue / Client / start_ml_evaluation_task_run
start_ml_evaluation_task_run#
- Glue.Client.start_ml_evaluation_task_run(**kwargs)#
Starts a task to estimate the quality of the transform.
When you provide label sets as examples of truth, Glue machine learning uses some of those examples to learn from them. The rest of the labels are used as a test to estimate quality.
Returns a unique identifier for the run. You can call
GetMLTaskRun
to get more information about the stats of theEvaluationTaskRun
.See also: AWS API Documentation
Request Syntax
response = client.start_ml_evaluation_task_run( TransformId='string' )
- Parameters:
TransformId (string) –
[REQUIRED]
The unique identifier of the machine learning transform.
- Return type:
dict
- Returns:
Response Syntax
{ 'TaskRunId': 'string' }
Response Structure
(dict) –
TaskRunId (string) –
The unique identifier associated with this run.
Exceptions
Glue.Client.exceptions.EntityNotFoundException
Glue.Client.exceptions.InvalidInputException
Glue.Client.exceptions.OperationTimeoutException
Glue.Client.exceptions.InternalServiceException
Glue.Client.exceptions.ConcurrentRunsExceededException
Glue.Client.exceptions.MLTransformNotReadyException