google_api_video_intelligence v0.23.0 API Reference
Modules
API calls for all endpoints tagged Operations
.
API calls for all endpoints tagged Projects
.
API calls for all endpoints tagged Videos
.
Handle Tesla connections for GoogleApi.VideoIntelligence.V1.
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Video annotation request.
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
A generic detected attribute represented by name in string format.
A generic detected landmark represented by name in string format and a 2D location.
Detected entity from video analysis.
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
Config for EXPLICIT_CONTENT_DETECTION.
Video frame level annotation results for explicit content.
Label annotation.
Config for LABEL_DETECTION.
Video frame level annotation results for label detection.
Video segment level annotation results for label detection.
Annotation corresponding to one detected, tracked and recognized logo class.
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
Annotations corresponding to one tracked object.
Config for OBJECT_TRACKING.
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
Config for SHOT_CHANGE_DETECTION.
Provides "hints" to the speech recognizer to favor specific words and phrases in the results.
Alternative hypotheses (a.k.a. n-best list).
A speech recognition result corresponding to a portion of the audio.
Config for SPEECH_TRANSCRIPTION.
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
Config for TEXT_DETECTION.
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
Video segment level annotation results for text detection.
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
A track of an object instance.
Annotation progress for a single video.
Annotation results for a single video.
Video context and/or feature-specific parameters.
Word-specific information for recognized words. Word information is only
included in the response when certain request parameters are set, such
as enable_word_time_offsets
.
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
A generic detected attribute represented by name in string format.
A generic detected landmark represented by name in string format and a 2D location.
Detected entity from video analysis.
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
Video frame level annotation results for explicit content.
Label annotation.
Video frame level annotation results for label detection.
Video segment level annotation results for label detection.
Annotation corresponding to one detected, tracked and recognized logo class.
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
Annotations corresponding to one tracked object.
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
Alternative hypotheses (a.k.a. n-best list).
A speech recognition result corresponding to a portion of the audio.
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
Video segment level annotation results for text detection.
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
A track of an object instance.
Annotation progress for a single video.
Annotation results for a single video.
Video segment.
Word-specific information for recognized words. Word information is only
included in the response when certain request parameters are set, such
as enable_word_time_offsets
.
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
A generic detected attribute represented by name in string format.
A generic detected landmark represented by name in string format and a 2D location.
Detected entity from video analysis.
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
Video frame level annotation results for explicit content.
Label annotation.
Video frame level annotation results for label detection.
Video segment level annotation results for label detection.
Annotation corresponding to one detected, tracked and recognized logo class.
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
Annotations corresponding to one tracked object.
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
Alternative hypotheses (a.k.a. n-best list).
A speech recognition result corresponding to a portion of the audio.
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
Video segment level annotation results for text detection.
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
A track of an object instance.
Annotation progress for a single video.
Annotation results for a single video.
Video segment.
Word-specific information for recognized words. Word information is only
included in the response when certain request parameters are set, such
as enable_word_time_offsets
.
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
A generic detected attribute represented by name in string format.
A generic detected landmark represented by name in string format and a 2D location.
Detected entity from video analysis.
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
Video frame level annotation results for explicit content.
Label annotation.
Video frame level annotation results for label detection.
Video segment level annotation results for label detection.
Annotation corresponding to one detected, tracked and recognized logo class.
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
Annotations corresponding to one tracked object.
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
Alternative hypotheses (a.k.a. n-best list).
A speech recognition result corresponding to a portion of the audio.
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
Video segment level annotation results for text detection.
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
A track of an object instance.
Annotation progress for a single video.
Annotation results for a single video.
Video segment.
Word-specific information for recognized words. Word information is only
included in the response when certain request parameters are set, such
as enable_word_time_offsets
.
Video annotation progress. Included in the metadata
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Video annotation response. Included in the response
field of the Operation
returned by the GetOperation
call of the google::longrunning::Operations
service.
Celebrity definition.
Celebrity recognition annotation per video.
The annotation result of a celebrity face track. RecognizedCelebrity field could be empty if the face track does not have any matched celebrities.
A generic detected attribute represented by name in string format.
A generic detected landmark represented by name in string format and a 2D location.
Detected entity from video analysis.
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
Video frame level annotation results for explicit content.
Face detection annotation.
Label annotation.
Video frame level annotation results for label detection.
Video segment level annotation results for label detection.
Annotation corresponding to one detected, tracked and recognized logo class.
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
Annotations corresponding to one tracked object.
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
Person detection annotation per video.
The recognized celebrity with confidence score.
Alternative hypotheses (a.k.a. n-best list).
A speech recognition result corresponding to a portion of the audio.
StreamingAnnotateVideoResponse
is the only message returned to the client
by StreamingAnnotateVideo
. A series of zero or more
StreamingAnnotateVideoResponse
messages are streamed back to the client.
Streaming annotation results corresponding to a portion of the video that is currently being processed.
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
Video segment level annotation results for text detection.
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
A track of an object instance.
Annotation progress for a single video.
Annotation results for a single video.
Video segment.
Word-specific information for recognized words. Word information is only
included in the response when certain request parameters are set, such
as enable_word_time_offsets
.
The request message for Operations.CancelOperation.
The response message for Operations.ListOperations.
This resource represents a long-running operation that is the result of a network API call.
A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance
The Status
type defines a logical error model that is suitable for
different programming environments, including REST APIs and RPC APIs. It is
used by gRPC. Each Status
message contains
three pieces of data: error code, error message, and error details.