Rekognition / Client / get_text_detection
get_text_detection#
- Rekognition.Client.get_text_detection(**kwargs)#
Gets the text detection results of a Amazon Rekognition Video analysis started by StartTextDetection.
Text detection with Amazon Rekognition Video is an asynchronous operation. You start text detection by calling StartTextDetection which returns a job identifier (
JobId
) When the text detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call toStartTextDetection
. To get the results of the text detection operation, first check that the status value published to the Amazon SNS topic isSUCCEEDED
. if so, callGetTextDetection
and pass the job identifier (JobId
) from the initial call ofStartLabelDetection
.GetTextDetection
returns an array of detected text (TextDetections
) sorted by the time the text was detected, up to 100 words per frame of video.Each element of the array includes the detected text, the precentage confidence in the acuracy of the detected text, the time the text was detected, bounding box information for where the text was located, and unique identifiers for words and their lines.
Use MaxResults parameter to limit the number of text detections returned. If there are more results than specified in
MaxResults
, the value ofNextToken
in the operation response contains a pagination token for getting the next set of results. To get the next page of results, callGetTextDetection
and populate theNextToken
request parameter with the token value returned from the previous call toGetTextDetection
.See also: AWS API Documentation
Request Syntax
response = client.get_text_detection( JobId='string', MaxResults=123, NextToken='string' )
- Parameters:
JobId (string) –
[REQUIRED]
Job identifier for the text detection operation for which you want results returned. You get the job identifer from an initial call to
StartTextDetection
.MaxResults (integer) – Maximum number of results to return per paginated call. The largest value you can specify is 1000.
NextToken (string) – If the previous response was incomplete (because there are more labels to retrieve), Amazon Rekognition Video returns a pagination token in the response. You can use this pagination token to retrieve the next set of text.
- Return type:
dict
- Returns:
Response Syntax
{ 'JobStatus': 'IN_PROGRESS'|'SUCCEEDED'|'FAILED', 'StatusMessage': 'string', 'VideoMetadata': { 'Codec': 'string', 'DurationMillis': 123, 'Format': 'string', 'FrameRate': ..., 'FrameHeight': 123, 'FrameWidth': 123, 'ColorRange': 'FULL'|'LIMITED' }, 'TextDetections': [ { 'Timestamp': 123, 'TextDetection': { 'DetectedText': 'string', 'Type': 'LINE'|'WORD', 'Id': 123, 'ParentId': 123, 'Confidence': ..., 'Geometry': { 'BoundingBox': { 'Width': ..., 'Height': ..., 'Left': ..., 'Top': ... }, 'Polygon': [ { 'X': ..., 'Y': ... }, ] } } }, ], 'NextToken': 'string', 'TextModelVersion': 'string', 'JobId': 'string', 'Video': { 'S3Object': { 'Bucket': 'string', 'Name': 'string', 'Version': 'string' } }, 'JobTag': 'string' }
Response Structure
(dict) –
JobStatus (string) –
Current status of the text detection job.
StatusMessage (string) –
If the job fails,
StatusMessage
provides a descriptive error message.VideoMetadata (dict) –
Information about a video that Amazon Rekognition analyzed.
Videometadata
is returned in every page of paginated responses from a Amazon Rekognition video operation.Codec (string) –
Type of compression used in the analyzed video.
DurationMillis (integer) –
Length of the video in milliseconds.
Format (string) –
Format of the analyzed video. Possible values are MP4, MOV and AVI.
FrameRate (float) –
Number of frames per second in the video.
FrameHeight (integer) –
Vertical pixel dimension of the video.
FrameWidth (integer) –
Horizontal pixel dimension of the video.
ColorRange (string) –
A description of the range of luminance values in a video, either LIMITED (16 to 235) or FULL (0 to 255).
TextDetections (list) –
An array of text detected in the video. Each element contains the detected text, the time in milliseconds from the start of the video that the text was detected, and where it was detected on the screen.
(dict) –
Information about text detected in a video. Incudes the detected text, the time in milliseconds from the start of the video that the text was detected, and where it was detected on the screen.
Timestamp (integer) –
The time, in milliseconds from the start of the video, that the text was detected. Note that
Timestamp
is not guaranteed to be accurate to the individual frame where the text first appears.TextDetection (dict) –
Details about text detected in a video.
DetectedText (string) –
The word or line of text recognized by Amazon Rekognition.
Type (string) –
The type of text that was detected.
Id (integer) –
The identifier for the detected text. The identifier is only unique for a single call to
DetectText
.ParentId (integer) –
The Parent identifier for the detected text identified by the value of
ID
. If the type of detected text isLINE
, the value ofParentId
isNull
.Confidence (float) –
The confidence that Amazon Rekognition has in the accuracy of the detected text and the accuracy of the geometry points around the detected text.
Geometry (dict) –
The location of the detected text on the image. Includes an axis aligned coarse bounding box surrounding the text and a finer grain polygon for more accurate spatial information.
BoundingBox (dict) –
An axis-aligned coarse representation of the detected item’s location on the image.
Width (float) –
Width of the bounding box as a ratio of the overall image width.
Height (float) –
Height of the bounding box as a ratio of the overall image height.
Left (float) –
Left coordinate of the bounding box as a ratio of overall image width.
Top (float) –
Top coordinate of the bounding box as a ratio of overall image height.
Polygon (list) –
Within the bounding box, a fine-grained polygon around the detected item.
(dict) –
The X and Y coordinates of a point on an image or video frame. The X and Y values are ratios of the overall image size or video resolution. For example, if an input image is 700x200 and the values are X=0.5 and Y=0.25, then the point is at the (350,50) pixel coordinate on the image.
An array of
Point
objects makes up aPolygon
. APolygon
is returned by DetectText and by DetectCustomLabelsPolygon
represents a fine-grained polygon around a detected item. For more information, see Geometry in the Amazon Rekognition Developer Guide.X (float) –
The value of the X coordinate for a point on a
Polygon
.Y (float) –
The value of the Y coordinate for a point on a
Polygon
.
NextToken (string) –
If the response is truncated, Amazon Rekognition Video returns this token that you can use in the subsequent request to retrieve the next set of text.
TextModelVersion (string) –
Version number of the text detection model that was used to detect text.
JobId (string) –
Job identifier for the text detection operation for which you want to obtain results. The job identifer is returned by an initial call to StartTextDetection.
Video (dict) –
Video file stored in an Amazon S3 bucket. Amazon Rekognition video start operations such as StartLabelDetection use
Video
to specify a video for analysis. The supported file formats are .mp4, .mov and .avi.S3Object (dict) –
The Amazon S3 bucket name and file name for the video.
Bucket (string) –
Name of the S3 bucket.
Name (string) –
S3 object key name.
Version (string) –
If the bucket is versioning enabled, you can specify the object version.
JobTag (string) –
A job identifier specified in the call to StartTextDetection and returned in the job completion notification sent to your Amazon Simple Notification Service topic.
Exceptions