From b2eff01c7fd2c90292a094fbdfee3000beff87ad Mon Sep 17 00:00:00 2001 From: Google APIs Date: Thu, 31 Jan 2019 00:36:50 +0000 Subject: [PATCH] Autogenerated update (2019-01-31) Update: - cloudasset_v1beta1 - cloudresourcemanager_v1 - cloudresourcemanager_v1beta1 - dlp_v2 - iam_v1 - storagetransfer_v1 - youtube_v3 --- api_names_out.yaml | 400 +++ generated/google/apis/cloudasset_v1beta1.rb | 2 +- .../google/apis/cloudasset_v1beta1/service.rb | 87 +- .../google/apis/cloudresourcemanager_v1.rb | 2 +- .../apis/cloudresourcemanager_v1/service.rb | 4 +- .../apis/cloudresourcemanager_v1beta1.rb | 2 +- .../cloudresourcemanager_v1beta1/service.rb | 4 +- generated/google/apis/dlp_v2.rb | 2 +- generated/google/apis/dlp_v2/classes.rb | 28 + .../google/apis/dlp_v2/representations.rb | 15 + generated/google/apis/dlp_v2/service.rb | 24 +- generated/google/apis/iam_v1.rb | 2 +- generated/google/apis/iam_v1/service.rb | 8 + generated/google/apis/storagetransfer_v1.rb | 2 +- .../google/apis/storagetransfer_v1/classes.rb | 2 + .../apis/videointelligence_v1p2beta1.rb | 35 + .../videointelligence_v1p2beta1/classes.rb | 2872 +++++++++++++++++ .../representations.rb | 1221 +++++++ .../videointelligence_v1p2beta1/service.rb | 93 + generated/google/apis/youtube_v3.rb | 2 +- generated/google/apis/youtube_v3/classes.rb | 113 +- .../google/apis/youtube_v3/representations.rb | 39 +- 22 files changed, 4929 insertions(+), 30 deletions(-) create mode 100644 generated/google/apis/videointelligence_v1p2beta1.rb create mode 100644 generated/google/apis/videointelligence_v1p2beta1/classes.rb create mode 100644 generated/google/apis/videointelligence_v1p2beta1/representations.rb create mode 100644 generated/google/apis/videointelligence_v1p2beta1/service.rb diff --git a/api_names_out.yaml b/api_names_out.yaml index 6bfdec739..72492d035 100644 --- a/api_names_out.yaml +++ b/api_names_out.yaml @@ -14737,6 +14737,10 @@ "/cloudasset:v1beta1/cloudasset.batchGetAssetsHistory/readTimeWindow.startTime": read_time_window_start_time "/cloudasset:v1beta1/cloudasset.exportAssets": export_assets "/cloudasset:v1beta1/cloudasset.exportAssets/parent": parent +"/cloudasset:v1beta1/cloudasset.folders.exportAssets": export_folder_assets +"/cloudasset:v1beta1/cloudasset.folders.exportAssets/parent": parent +"/cloudasset:v1beta1/cloudasset.folders.operations.get": get_folder_operation +"/cloudasset:v1beta1/cloudasset.folders.operations.get/name": name "/cloudasset:v1beta1/cloudasset.operations.get": get_operation "/cloudasset:v1beta1/cloudasset.operations.get/name": name "/cloudasset:v1beta1/cloudasset.organizations.batchGetAssetsHistory": batch_organization_get_assets_history @@ -63699,6 +63703,7 @@ "/discovery:v1/quotaUser": quota_user "/discovery:v1/userIp": user_ip "/dlp:v2/GooglePrivacyDlpV2Action": google_privacy_dlp_v2_action +"/dlp:v2/GooglePrivacyDlpV2Action/jobNotificationEmails": job_notification_emails "/dlp:v2/GooglePrivacyDlpV2Action/pubSub": pub_sub "/dlp:v2/GooglePrivacyDlpV2Action/publishSummaryToCscc": publish_summary_to_cscc "/dlp:v2/GooglePrivacyDlpV2Action/saveFindings": save_findings @@ -64046,6 +64051,7 @@ "/dlp:v2/GooglePrivacyDlpV2InspectionRuleSet/infoTypes/info_type": info_type "/dlp:v2/GooglePrivacyDlpV2InspectionRuleSet/rules": rules "/dlp:v2/GooglePrivacyDlpV2InspectionRuleSet/rules/rule": rule +"/dlp:v2/GooglePrivacyDlpV2JobNotificationEmails": google_privacy_dlp_v2_job_notification_emails "/dlp:v2/GooglePrivacyDlpV2JobTrigger": google_privacy_dlp_v2_job_trigger "/dlp:v2/GooglePrivacyDlpV2JobTrigger/createTime": create_time "/dlp:v2/GooglePrivacyDlpV2JobTrigger/description": description @@ -64249,6 +64255,8 @@ "/dlp:v2/GooglePrivacyDlpV2RedactImageResponse/inspectResult": inspect_result "/dlp:v2/GooglePrivacyDlpV2RedactImageResponse/redactedImage": redacted_image "/dlp:v2/GooglePrivacyDlpV2Regex": google_privacy_dlp_v2_regex +"/dlp:v2/GooglePrivacyDlpV2Regex/groupIndexes": group_indexes +"/dlp:v2/GooglePrivacyDlpV2Regex/groupIndexes/group_index": group_index "/dlp:v2/GooglePrivacyDlpV2Regex/pattern": pattern "/dlp:v2/GooglePrivacyDlpV2ReidentifyContentRequest": google_privacy_dlp_v2_reidentify_content_request "/dlp:v2/GooglePrivacyDlpV2ReidentifyContentRequest/inspectConfig": inspect_config @@ -64691,6 +64699,7 @@ "/dlp:v2/dlp.projects.jobTriggers.get": get_project_job_trigger "/dlp:v2/dlp.projects.jobTriggers.get/name": name "/dlp:v2/dlp.projects.jobTriggers.list": list_project_job_triggers +"/dlp:v2/dlp.projects.jobTriggers.list/filter": filter "/dlp:v2/dlp.projects.jobTriggers.list/orderBy": order_by "/dlp:v2/dlp.projects.jobTriggers.list/pageSize": page_size "/dlp:v2/dlp.projects.jobTriggers.list/pageToken": page_token @@ -98870,6 +98879,384 @@ "/videointelligence:v1p1beta1/key": key "/videointelligence:v1p1beta1/quotaUser": quota_user "/videointelligence:v1p1beta1/videointelligence.videos.annotate": annotate_video +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoProgress": google_cloud_videointelligence_v1_annotate_video_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoProgress/annotationProgress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoProgress/annotationProgress/annotation_progress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoResponse": google_cloud_videointelligence_v1_annotate_video_response +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoResponse/annotationResults": annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_AnnotateVideoResponse/annotationResults/annotation_result": annotation_result +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_Entity": google_cloud_videointelligence_v1_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_Entity/description": description +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_Entity/entityId": entity_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_Entity/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentAnnotation": google_cloud_videointelligence_v1_explicit_content_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentFrame": google_cloud_videointelligence_v1_explicit_content_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentFrame/pornographyLikelihood": pornography_likelihood +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_ExplicitContentFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation": google_cloud_videointelligence_v1_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/categoryEntities": category_entities +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/categoryEntities/category_entity": category_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/entity": entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelAnnotation/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelFrame": google_cloud_videointelligence_v1_label_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelFrame/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelSegment": google_cloud_videointelligence_v1_label_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelSegment/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_LabelSegment/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechRecognitionAlternative": google_cloud_videointelligence_v1_speech_recognition_alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechRecognitionAlternative/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechRecognitionAlternative/transcript": transcript +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechRecognitionAlternative/words": words +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechRecognitionAlternative/words/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechTranscription": google_cloud_videointelligence_v1_speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechTranscription/alternatives": alternatives +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechTranscription/alternatives/alternative": alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_SpeechTranscription/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationProgress": google_cloud_videointelligence_v1_video_annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationProgress/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationProgress/progressPercent": progress_percent +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationProgress/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationProgress/updateTime": update_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults": google_cloud_videointelligence_v1_video_annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/error": error +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/explicitAnnotation": explicit_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/frameLabelAnnotations": frame_label_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/frameLabelAnnotations/frame_label_annotation": frame_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/segmentLabelAnnotations": segment_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/segmentLabelAnnotations/segment_label_annotation" +: segment_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/shotAnnotations": shot_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/shotAnnotations/shot_annotation": shot_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/shotLabelAnnotations": shot_label_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/shotLabelAnnotations/shot_label_annotation": shot_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/speechTranscriptions": speech_transcriptions +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoAnnotationResults/speechTranscriptions/speech_transcription": speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoSegment": google_cloud_videointelligence_v1_video_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoSegment/endTimeOffset": end_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_VideoSegment/startTimeOffset": start_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo": google_cloud_videointelligence_v1_word_info +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo/endTime": end_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo/speakerTag": speaker_tag +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1_WordInfo/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoProgress": google_cloud_videointelligence_v1beta2_annotate_video_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoProgress/annotationProgress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoProgress/annotationProgress/annotation_progress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoResponse": google_cloud_videointelligence_v1beta2_annotate_video_response +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoResponse/annotationResults": annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_AnnotateVideoResponse/annotationResults/annotation_result": annotation_result +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_Entity": google_cloud_videointelligence_v1beta2_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_Entity/description": description +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_Entity/entityId": entity_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_Entity/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentAnnotation": google_cloud_videointelligence_v1beta2_explicit_content_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentFrame": google_cloud_videointelligence_v1beta2_explicit_content_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentFrame/pornographyLikelihood": pornography_likelihood +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_ExplicitContentFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation": google_cloud_videointelligence_v1beta2_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/categoryEntities": category_entities +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/categoryEntities/category_entity": category_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/entity": entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelAnnotation/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelFrame": google_cloud_videointelligence_v1beta2_label_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelFrame/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelSegment": google_cloud_videointelligence_v1beta2_label_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelSegment/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_LabelSegment/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechRecognitionAlternative": google_cloud_videointelligence_v1beta2_speech_recognition_alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechRecognitionAlternative/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechRecognitionAlternative/transcript": transcript +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechRecognitionAlternative/words": words +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechRecognitionAlternative/words/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechTranscription": google_cloud_videointelligence_v1beta2_speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechTranscription/alternatives": alternatives +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechTranscription/alternatives/alternative": alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_SpeechTranscription/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationProgress": google_cloud_videointelligence_v1beta2_video_annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationProgress/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationProgress/progressPercent": progress_percent +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationProgress/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationProgress/updateTime": update_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults": google_cloud_videointelligence_v1beta2_video_annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/error": error +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/explicitAnnotation": explicit_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/frameLabelAnnotations": frame_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/frameLabelAnnotations/frame_label_annotation" +: frame_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/segmentLabelAnnotations": segment_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/segmentLabelAnnotations/segment_label_annotation" +: segment_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/shotAnnotations": shot_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/shotAnnotations/shot_annotation": shot_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/shotLabelAnnotations": shot_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/shotLabelAnnotations/shot_label_annotation" +: shot_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/speechTranscriptions": speech_transcriptions +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoAnnotationResults/speechTranscriptions/speech_transcription" +: speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoSegment": google_cloud_videointelligence_v1beta2_video_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoSegment/endTimeOffset": end_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_VideoSegment/startTimeOffset": start_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo": google_cloud_videointelligence_v1beta2_word_info +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo/endTime": end_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo/speakerTag": speaker_tag +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1beta2_WordInfo/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoProgress": google_cloud_videointelligence_v1p1beta1_annotate_video_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoProgress/annotationProgress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoProgress/annotationProgress/annotation_progress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoResponse": google_cloud_videointelligence_v1p1beta1_annotate_video_response +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoResponse/annotationResults": annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_AnnotateVideoResponse/annotationResults/annotation_result": annotation_result +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_Entity": google_cloud_videointelligence_v1p1beta1_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_Entity/description": description +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_Entity/entityId": entity_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_Entity/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentAnnotation": google_cloud_videointelligence_v1p1beta1_explicit_content_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentFrame": google_cloud_videointelligence_v1p1beta1_explicit_content_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentFrame/pornographyLikelihood": pornography_likelihood +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_ExplicitContentFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation": google_cloud_videointelligence_v1p1beta1_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/categoryEntities": category_entities +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/categoryEntities/category_entity": category_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/entity": entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelAnnotation/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelFrame": google_cloud_videointelligence_v1p1beta1_label_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelFrame/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelSegment": google_cloud_videointelligence_v1p1beta1_label_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelSegment/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_LabelSegment/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechRecognitionAlternative": google_cloud_videointelligence_v1p1beta1_speech_recognition_alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechRecognitionAlternative/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechRecognitionAlternative/transcript": transcript +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechRecognitionAlternative/words": words +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechRecognitionAlternative/words/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechTranscription": google_cloud_videointelligence_v1p1beta1_speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechTranscription/alternatives": alternatives +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechTranscription/alternatives/alternative": alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_SpeechTranscription/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationProgress": google_cloud_videointelligence_v1p1beta1_video_annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationProgress/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationProgress/progressPercent": progress_percent +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationProgress/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationProgress/updateTime": update_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults": google_cloud_videointelligence_v1p1beta1_video_annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/error": error +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/explicitAnnotation": explicit_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/frameLabelAnnotations": frame_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/frameLabelAnnotations/frame_label_annotation" +: frame_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/segmentLabelAnnotations": segment_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/segmentLabelAnnotations/segment_label_annotation" +: segment_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/shotAnnotations": shot_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/shotAnnotations/shot_annotation": shot_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/shotLabelAnnotations": shot_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/shotLabelAnnotations/shot_label_annotation" +: shot_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/speechTranscriptions": speech_transcriptions +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoAnnotationResults/speechTranscriptions/speech_transcription" +: speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoSegment": google_cloud_videointelligence_v1p1beta1_video_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoSegment/endTimeOffset": end_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_VideoSegment/startTimeOffset": start_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo": google_cloud_videointelligence_v1p1beta1_word_info +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo/endTime": end_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo/speakerTag": speaker_tag +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p1beta1_WordInfo/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoProgress": google_cloud_videointelligence_v1p2beta1_annotate_video_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoProgress/annotationProgress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoProgress/annotationProgress/annotation_progress": annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest": google_cloud_videointelligence_v1p2beta1_annotate_video_request +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/features": features +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/features/feature": feature +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/inputContent": input_content +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/locationId": location_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/outputUri": output_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoRequest/videoContext": video_context +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoResponse": google_cloud_videointelligence_v1p2beta1_annotate_video_response +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoResponse/annotationResults": annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_AnnotateVideoResponse/annotationResults/annotation_result": annotation_result +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_Entity": google_cloud_videointelligence_v1p2beta1_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_Entity/description": description +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_Entity/entityId": entity_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_Entity/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentAnnotation": google_cloud_videointelligence_v1p2beta1_explicit_content_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentDetectionConfig": google_cloud_videointelligence_v1p2beta1_explicit_content_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentDetectionConfig/model": model +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentFrame": google_cloud_videointelligence_v1p2beta1_explicit_content_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentFrame/pornographyLikelihood": pornography_likelihood +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ExplicitContentFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation": google_cloud_videointelligence_v1p2beta1_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/categoryEntities": category_entities +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/categoryEntities/category_entity": category_entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/entity": entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelAnnotation/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelDetectionConfig": google_cloud_videointelligence_v1p2beta1_label_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelDetectionConfig/labelDetectionMode": label_detection_mode +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelDetectionConfig/model": model +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelDetectionConfig/stationaryCamera": stationary_camera +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelFrame": google_cloud_videointelligence_v1p2beta1_label_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelFrame/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelSegment": google_cloud_videointelligence_v1p2beta1_label_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelSegment/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_LabelSegment/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingBox": google_cloud_videointelligence_v1p2beta1_normalized_bounding_box +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingBox/bottom": bottom +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingBox/left": left +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingBox/right": right +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingBox/top": top +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingPoly": google_cloud_videointelligence_v1p2beta1_normalized_bounding_poly +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingPoly/vertices": vertices +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedBoundingPoly/vertices/vertex": vertex +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedVertex": google_cloud_videointelligence_v1p2beta1_normalized_vertex +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedVertex/x": x +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_NormalizedVertex/y": y +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation": google_cloud_videointelligence_v1p2beta1_object_tracking_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/entity": entity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingAnnotation/trackId": track_id +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingFrame": google_cloud_videointelligence_v1p2beta1_object_tracking_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingFrame/normalizedBoundingBox": normalized_bounding_box +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ObjectTrackingFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ShotChangeDetectionConfig": google_cloud_videointelligence_v1p2beta1_shot_change_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_ShotChangeDetectionConfig/model": model +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechContext": google_cloud_videointelligence_v1p2beta1_speech_context +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechContext/phrases": phrases +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechContext/phrases/phrase": phrase +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechRecognitionAlternative": google_cloud_videointelligence_v1p2beta1_speech_recognition_alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechRecognitionAlternative/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechRecognitionAlternative/transcript": transcript +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechRecognitionAlternative/words": words +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechRecognitionAlternative/words/word": word +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscription": google_cloud_videointelligence_v1p2beta1_speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscription/alternatives": alternatives +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscription/alternatives/alternative": alternative +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscription/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig": google_cloud_videointelligence_v1p2beta1_speech_transcription_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/audioTracks": audio_tracks +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/audioTracks/audio_track": audio_track +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/diarizationSpeakerCount": diarization_speaker_count +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/enableAutomaticPunctuation": enable_automatic_punctuation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/enableSpeakerDiarization": enable_speaker_diarization +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/enableWordConfidence": enable_word_confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/filterProfanity": filter_profanity +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/languageCode": language_code +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/maxAlternatives": max_alternatives +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/speechContexts": speech_contexts +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_SpeechTranscriptionConfig/speechContexts/speech_context": speech_context +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextAnnotation": google_cloud_videointelligence_v1p2beta1_text_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextAnnotation/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextAnnotation/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextAnnotation/text": text +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextDetectionConfig": google_cloud_videointelligence_v1p2beta1_text_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextDetectionConfig/languageHints": language_hints +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextDetectionConfig/languageHints/language_hint": language_hint +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextFrame": google_cloud_videointelligence_v1p2beta1_text_frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextFrame/rotatedBoundingBox": rotated_bounding_box +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextFrame/timeOffset": time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextSegment": google_cloud_videointelligence_v1p2beta1_text_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextSegment/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextSegment/frames": frames +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextSegment/frames/frame": frame +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_TextSegment/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationProgress": google_cloud_videointelligence_v1p2beta1_video_annotation_progress +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationProgress/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationProgress/progressPercent": progress_percent +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationProgress/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationProgress/updateTime": update_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults": google_cloud_videointelligence_v1p2beta1_video_annotation_results +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/error": error +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/explicitAnnotation": explicit_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/frameLabelAnnotations": frame_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/frameLabelAnnotations/frame_label_annotation" +: frame_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/inputUri": input_uri +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/objectAnnotations": object_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/objectAnnotations/object_annotation": object_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/segmentLabelAnnotations": segment_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/segmentLabelAnnotations/segment_label_annotation" +: segment_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/shotAnnotations": shot_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/shotAnnotations/shot_annotation": shot_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/shotLabelAnnotations": shot_label_annotations +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/shotLabelAnnotations/shot_label_annotation" +: shot_label_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/speechTranscriptions": speech_transcriptions +? "/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/speechTranscriptions/speech_transcription" +: speech_transcription +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/textAnnotations": text_annotations +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoAnnotationResults/textAnnotations/text_annotation": text_annotation +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext": google_cloud_videointelligence_v1p2beta1_video_context +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/explicitContentDetectionConfig": explicit_content_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/labelDetectionConfig": label_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/segments": segments +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/segments/segment": segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/shotChangeDetectionConfig": shot_change_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/speechTranscriptionConfig": speech_transcription_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoContext/textDetectionConfig": text_detection_config +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoSegment": google_cloud_videointelligence_v1p2beta1_video_segment +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoSegment/endTimeOffset": end_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_VideoSegment/startTimeOffset": start_time_offset +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo": google_cloud_videointelligence_v1p2beta1_word_info +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo/confidence": confidence +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo/endTime": end_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo/speakerTag": speaker_tag +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo/startTime": start_time +"/videointelligence:v1p2beta1/GoogleCloudVideointelligenceV1p2beta1_WordInfo/word": word +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation": google_longrunning_operation +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/done": done +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/error": error +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/metadata": metadata +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/metadata/metadatum": metadatum +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/name": name +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/response": response +"/videointelligence:v1p2beta1/GoogleLongrunning_Operation/response/response": response +"/videointelligence:v1p2beta1/GoogleRpc_Status": google_rpc_status +"/videointelligence:v1p2beta1/GoogleRpc_Status/code": code +"/videointelligence:v1p2beta1/GoogleRpc_Status/details": details +"/videointelligence:v1p2beta1/GoogleRpc_Status/details/detail": detail +"/videointelligence:v1p2beta1/GoogleRpc_Status/details/detail/detail": detail +"/videointelligence:v1p2beta1/GoogleRpc_Status/message": message +"/videointelligence:v1p2beta1/fields": fields +"/videointelligence:v1p2beta1/key": key +"/videointelligence:v1p2beta1/quotaUser": quota_user +"/videointelligence:v1p2beta1/videointelligence.videos.annotate": annotate_video "/vision:v1/AddProductToProductSetRequest": add_product_to_product_set_request "/vision:v1/AddProductToProductSetRequest/product": product "/vision:v1/AnnotateFileResponse": annotate_file_response @@ -103165,6 +103552,7 @@ "/youtube:v3/LiveChatMessageSnippet/pollVotedDetails": poll_voted_details "/youtube:v3/LiveChatMessageSnippet/publishedAt": published_at "/youtube:v3/LiveChatMessageSnippet/superChatDetails": super_chat_details +"/youtube:v3/LiveChatMessageSnippet/superStickerDetails": super_sticker_details "/youtube:v3/LiveChatMessageSnippet/textMessageDetails": text_message_details "/youtube:v3/LiveChatMessageSnippet/type": type "/youtube:v3/LiveChatMessageSnippet/userBannedDetails": user_banned_details @@ -103211,6 +103599,12 @@ "/youtube:v3/LiveChatSuperChatDetails/currency": currency "/youtube:v3/LiveChatSuperChatDetails/tier": tier "/youtube:v3/LiveChatSuperChatDetails/userComment": user_comment +"/youtube:v3/LiveChatSuperStickerDetails": live_chat_super_sticker_details +"/youtube:v3/LiveChatSuperStickerDetails/amountDisplayString": amount_display_string +"/youtube:v3/LiveChatSuperStickerDetails/amountMicros": amount_micros +"/youtube:v3/LiveChatSuperStickerDetails/currency": currency +"/youtube:v3/LiveChatSuperStickerDetails/superStickerMetadata": super_sticker_metadata +"/youtube:v3/LiveChatSuperStickerDetails/tier": tier "/youtube:v3/LiveChatTextMessageDetails": live_chat_text_message_details "/youtube:v3/LiveChatTextMessageDetails/messageText": message_text "/youtube:v3/LiveChatUserBannedMessageDetails": live_chat_user_banned_message_details @@ -103476,9 +103870,15 @@ "/youtube:v3/SuperChatEventSnippet/currency": currency "/youtube:v3/SuperChatEventSnippet/displayString": display_string "/youtube:v3/SuperChatEventSnippet/isSuperChatForGood": is_super_chat_for_good +"/youtube:v3/SuperChatEventSnippet/isSuperStickerEvent": is_super_sticker_event "/youtube:v3/SuperChatEventSnippet/messageType": message_type "/youtube:v3/SuperChatEventSnippet/nonprofit": nonprofit +"/youtube:v3/SuperChatEventSnippet/superStickerMetadata": super_sticker_metadata "/youtube:v3/SuperChatEventSnippet/supporterDetails": supporter_details +"/youtube:v3/SuperStickerMetadata": super_sticker_metadata +"/youtube:v3/SuperStickerMetadata/altText": alt_text +"/youtube:v3/SuperStickerMetadata/altTextLanguage": alt_text_language +"/youtube:v3/SuperStickerMetadata/stickerId": sticker_id "/youtube:v3/Thumbnail": thumbnail "/youtube:v3/Thumbnail/height": height "/youtube:v3/Thumbnail/url": url diff --git a/generated/google/apis/cloudasset_v1beta1.rb b/generated/google/apis/cloudasset_v1beta1.rb index 1fc04d327..378c29844 100644 --- a/generated/google/apis/cloudasset_v1beta1.rb +++ b/generated/google/apis/cloudasset_v1beta1.rb @@ -25,7 +25,7 @@ module Google # @see https://console.cloud.google.com/apis/api/cloudasset.googleapis.com/overview module CloudassetV1beta1 VERSION = 'V1beta1' - REVISION = '20181204' + REVISION = '20190128' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/cloudasset_v1beta1/service.rb b/generated/google/apis/cloudasset_v1beta1/service.rb index 79d619b5f..6358e9f3b 100644 --- a/generated/google/apis/cloudasset_v1beta1/service.rb +++ b/generated/google/apis/cloudasset_v1beta1/service.rb @@ -47,11 +47,84 @@ module Google @batch_path = 'batch' end + # Exports assets with time and resource types to a given Cloud Storage + # location. The output format is newline-delimited JSON. + # This API implements the google.longrunning.Operation API allowing you + # to keep track of the export. + # @param [String] parent + # Required. The relative name of the root asset. This can only be an + # organization number (such as "organizations/123"), a project ID (such as + # "projects/my-project-id"), a project number (such as "projects/12345"), or + # a folder number (such as "folders/123"). + # @param [Google::Apis::CloudassetV1beta1::ExportAssetsRequest] export_assets_request_object + # @param [String] fields + # Selector specifying which fields to include in a partial response. + # @param [String] quota_user + # Available to use for quota purposes for server-side applications. Can be any + # arbitrary string assigned to a user, but should not exceed 40 characters. + # @param [Google::Apis::RequestOptions] options + # Request-specific options + # + # @yield [result, err] Result & error if block supplied + # @yieldparam result [Google::Apis::CloudassetV1beta1::Operation] parsed result object + # @yieldparam err [StandardError] error object if request failed + # + # @return [Google::Apis::CloudassetV1beta1::Operation] + # + # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried + # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification + # @raise [Google::Apis::AuthorizationError] Authorization is required + def export_folder_assets(parent, export_assets_request_object = nil, fields: nil, quota_user: nil, options: nil, &block) + command = make_simple_command(:post, 'v1beta1/{+parent}:exportAssets', options) + command.request_representation = Google::Apis::CloudassetV1beta1::ExportAssetsRequest::Representation + command.request_object = export_assets_request_object + command.response_representation = Google::Apis::CloudassetV1beta1::Operation::Representation + command.response_class = Google::Apis::CloudassetV1beta1::Operation + command.params['parent'] = parent unless parent.nil? + command.query['fields'] = fields unless fields.nil? + command.query['quotaUser'] = quota_user unless quota_user.nil? + execute_or_queue_command(command, &block) + end + + # Gets the latest state of a long-running operation. Clients can use this + # method to poll the operation result at intervals as recommended by the API + # service. + # @param [String] name + # The name of the operation resource. + # @param [String] fields + # Selector specifying which fields to include in a partial response. + # @param [String] quota_user + # Available to use for quota purposes for server-side applications. Can be any + # arbitrary string assigned to a user, but should not exceed 40 characters. + # @param [Google::Apis::RequestOptions] options + # Request-specific options + # + # @yield [result, err] Result & error if block supplied + # @yieldparam result [Google::Apis::CloudassetV1beta1::Operation] parsed result object + # @yieldparam err [StandardError] error object if request failed + # + # @return [Google::Apis::CloudassetV1beta1::Operation] + # + # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried + # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification + # @raise [Google::Apis::AuthorizationError] Authorization is required + def get_folder_operation(name, fields: nil, quota_user: nil, options: nil, &block) + command = make_simple_command(:get, 'v1beta1/{+name}', options) + command.response_representation = Google::Apis::CloudassetV1beta1::Operation::Representation + command.response_class = Google::Apis::CloudassetV1beta1::Operation + command.params['name'] = name unless name.nil? + command.query['fields'] = fields unless fields.nil? + command.query['quotaUser'] = quota_user unless quota_user.nil? + execute_or_queue_command(command, &block) + end + # Batch gets the update history of assets that overlap a time window. # For RESOURCE content, this API outputs history with asset in both # non-delete or deleted status. # For IAM_POLICY content, this API outputs history when the asset and its # attached IAM POLICY both exist. This can create gaps in the output history. + # If a specified asset does not exist, this API returns an INVALID_ARGUMENT + # error. # @param [String] parent # Required. The relative name of the root asset. It can only be an # organization number (such as "organizations/123"), a project ID (such as @@ -109,9 +182,9 @@ module Google # to keep track of the export. # @param [String] parent # Required. The relative name of the root asset. This can only be an - # organization - # number (such as "organizations/123"), a project ID (such as - # "projects/my-project-id"), or a project number (such as "projects/12345"). + # organization number (such as "organizations/123"), a project ID (such as + # "projects/my-project-id"), a project number (such as "projects/12345"), or + # a folder number (such as "folders/123"). # @param [Google::Apis::CloudassetV1beta1::ExportAssetsRequest] export_assets_request_object # @param [String] fields # Selector specifying which fields to include in a partial response. @@ -179,6 +252,8 @@ module Google # non-delete or deleted status. # For IAM_POLICY content, this API outputs history when the asset and its # attached IAM POLICY both exist. This can create gaps in the output history. + # If a specified asset does not exist, this API returns an INVALID_ARGUMENT + # error. # @param [String] parent # Required. The relative name of the root asset. It can only be an # organization number (such as "organizations/123"), a project ID (such as @@ -236,9 +311,9 @@ module Google # to keep track of the export. # @param [String] parent # Required. The relative name of the root asset. This can only be an - # organization - # number (such as "organizations/123"), a project ID (such as - # "projects/my-project-id"), or a project number (such as "projects/12345"). + # organization number (such as "organizations/123"), a project ID (such as + # "projects/my-project-id"), a project number (such as "projects/12345"), or + # a folder number (such as "folders/123"). # @param [Google::Apis::CloudassetV1beta1::ExportAssetsRequest] export_assets_request_object # @param [String] fields # Selector specifying which fields to include in a partial response. diff --git a/generated/google/apis/cloudresourcemanager_v1.rb b/generated/google/apis/cloudresourcemanager_v1.rb index 3c6b2fa5b..e924f9de1 100644 --- a/generated/google/apis/cloudresourcemanager_v1.rb +++ b/generated/google/apis/cloudresourcemanager_v1.rb @@ -26,7 +26,7 @@ module Google # @see https://cloud.google.com/resource-manager module CloudresourcemanagerV1 VERSION = 'V1' - REVISION = '20181015' + REVISION = '20190128' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/cloudresourcemanager_v1/service.rb b/generated/google/apis/cloudresourcemanager_v1/service.rb index fd4a6cbf4..3a1dc8c86 100644 --- a/generated/google/apis/cloudresourcemanager_v1/service.rb +++ b/generated/google/apis/cloudresourcemanager_v1/service.rb @@ -465,7 +465,9 @@ module Google # Fetches an Organization resource identified by the specified resource name. # @param [String] name - # The resource name of the Organization to fetch, e.g. "organizations/1234". + # The resource name of the Organization to fetch. This is the organization's + # relative path in the API, formatted as "organizations/[organizationId]". + # For example, "organizations/1234". # @param [String] fields # Selector specifying which fields to include in a partial response. # @param [String] quota_user diff --git a/generated/google/apis/cloudresourcemanager_v1beta1.rb b/generated/google/apis/cloudresourcemanager_v1beta1.rb index 0f54df76f..c5c3ee7f1 100644 --- a/generated/google/apis/cloudresourcemanager_v1beta1.rb +++ b/generated/google/apis/cloudresourcemanager_v1beta1.rb @@ -26,7 +26,7 @@ module Google # @see https://cloud.google.com/resource-manager module CloudresourcemanagerV1beta1 VERSION = 'V1beta1' - REVISION = '20181015' + REVISION = '20190128' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/cloudresourcemanager_v1beta1/service.rb b/generated/google/apis/cloudresourcemanager_v1beta1/service.rb index 5c344c95d..9ad694160 100644 --- a/generated/google/apis/cloudresourcemanager_v1beta1/service.rb +++ b/generated/google/apis/cloudresourcemanager_v1beta1/service.rb @@ -50,7 +50,9 @@ module Google # Fetches an Organization resource identified by the specified resource name. # @param [String] name - # The resource name of the Organization to fetch, e.g. "organizations/1234". + # The resource name of the Organization to fetch. This is the organization's + # relative path in the API, formatted as "organizations/[organizationId]". + # For example, "organizations/1234". # @param [String] organization_id # The id of the Organization resource to fetch. # This field is deprecated and will be removed in v1. Use name instead. diff --git a/generated/google/apis/dlp_v2.rb b/generated/google/apis/dlp_v2.rb index d8ebcdad3..9ace59ec2 100644 --- a/generated/google/apis/dlp_v2.rb +++ b/generated/google/apis/dlp_v2.rb @@ -27,7 +27,7 @@ module Google # @see https://cloud.google.com/dlp/docs/ module DlpV2 VERSION = 'V2' - REVISION = '20181229' + REVISION = '20190129' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/dlp_v2/classes.rb b/generated/google/apis/dlp_v2/classes.rb index 497289119..807e198dd 100644 --- a/generated/google/apis/dlp_v2/classes.rb +++ b/generated/google/apis/dlp_v2/classes.rb @@ -27,6 +27,12 @@ module Google class GooglePrivacyDlpV2Action include Google::Apis::Core::Hashable + # Enable email notification to project owners and editors on jobs's + # completion/failure. + # Corresponds to the JSON property `jobNotificationEmails` + # @return [Google::Apis::DlpV2::GooglePrivacyDlpV2JobNotificationEmails] + attr_accessor :job_notification_emails + # Publish the results of a DlpJob to a pub sub channel. # Compatible with: Inspect, Risk # Corresponds to the JSON property `pubSub` @@ -61,6 +67,7 @@ module Google # Update properties of this object def update!(**args) + @job_notification_emails = args[:job_notification_emails] if args.key?(:job_notification_emails) @pub_sub = args[:pub_sub] if args.key?(:pub_sub) @publish_summary_to_cscc = args[:publish_summary_to_cscc] if args.key?(:publish_summary_to_cscc) @save_findings = args[:save_findings] if args.key?(:save_findings) @@ -3057,6 +3064,20 @@ module Google end end + # Enable email notification to project owners and editors on jobs's + # completion/failure. + class GooglePrivacyDlpV2JobNotificationEmails + include Google::Apis::Core::Hashable + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + end + end + # Contains a configuration to make dlp api calls on a repeating basis. # See https://cloud.google.com/dlp/docs/concepts-job-triggers to learn more. class GooglePrivacyDlpV2JobTrigger @@ -4658,6 +4679,12 @@ module Google class GooglePrivacyDlpV2Regex include Google::Apis::Core::Hashable + # The index of the submatch to extract as findings. When not + # specified, the entire match is returned. No more than 3 may be included. + # Corresponds to the JSON property `groupIndexes` + # @return [Array] + attr_accessor :group_indexes + # Pattern defining the regular expression. Its syntax # (https://github.com/google/re2/wiki/Syntax) can be found under the # google/re2 repository on GitHub. @@ -4671,6 +4698,7 @@ module Google # Update properties of this object def update!(**args) + @group_indexes = args[:group_indexes] if args.key?(:group_indexes) @pattern = args[:pattern] if args.key?(:pattern) end end diff --git a/generated/google/apis/dlp_v2/representations.rb b/generated/google/apis/dlp_v2/representations.rb index 047dc31a2..ca99ebac2 100644 --- a/generated/google/apis/dlp_v2/representations.rb +++ b/generated/google/apis/dlp_v2/representations.rb @@ -502,6 +502,12 @@ module Google include Google::Apis::Core::JsonObjectSupport end + class GooglePrivacyDlpV2JobNotificationEmails + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + class GooglePrivacyDlpV2JobTrigger class Representation < Google::Apis::Core::JsonRepresentation; end @@ -1039,6 +1045,8 @@ module Google class GooglePrivacyDlpV2Action # @private class Representation < Google::Apis::Core::JsonRepresentation + property :job_notification_emails, as: 'jobNotificationEmails', class: Google::Apis::DlpV2::GooglePrivacyDlpV2JobNotificationEmails, decorator: Google::Apis::DlpV2::GooglePrivacyDlpV2JobNotificationEmails::Representation + property :pub_sub, as: 'pubSub', class: Google::Apis::DlpV2::GooglePrivacyDlpV2PublishToPubSub, decorator: Google::Apis::DlpV2::GooglePrivacyDlpV2PublishToPubSub::Representation property :publish_summary_to_cscc, as: 'publishSummaryToCscc', class: Google::Apis::DlpV2::GooglePrivacyDlpV2PublishSummaryToCscc, decorator: Google::Apis::DlpV2::GooglePrivacyDlpV2PublishSummaryToCscc::Representation @@ -1880,6 +1888,12 @@ module Google end end + class GooglePrivacyDlpV2JobNotificationEmails + # @private + class Representation < Google::Apis::Core::JsonRepresentation + end + end + class GooglePrivacyDlpV2JobTrigger # @private class Representation < Google::Apis::Core::JsonRepresentation @@ -2369,6 +2383,7 @@ module Google class GooglePrivacyDlpV2Regex # @private class Representation < Google::Apis::Core::JsonRepresentation + collection :group_indexes, as: 'groupIndexes' property :pattern, as: 'pattern' end end diff --git a/generated/google/apis/dlp_v2/service.rb b/generated/google/apis/dlp_v2/service.rb index be7d40c40..a61c3f5e6 100644 --- a/generated/google/apis/dlp_v2/service.rb +++ b/generated/google/apis/dlp_v2/service.rb @@ -1524,6 +1524,26 @@ module Google # See https://cloud.google.com/dlp/docs/creating-job-triggers to learn more. # @param [String] parent # The parent resource name, for example `projects/my-project-id`. + # @param [String] filter + # Optional. Allows filtering. + # Supported syntax: + # * Filter expressions are made up of one or more restrictions. + # * Restrictions can be combined by `AND` or `OR` logical operators. A + # sequence of restrictions implicitly uses `AND`. + # * A restriction has the form of ` `. + # * Supported fields/values for inspect jobs: + # - `status` - HEALTHY|PAUSED|CANCELLED + # - `inspected_storage` - DATASTORE|CLOUD_STORAGE|BIGQUERY + # - 'last_run_time` - RFC 3339 formatted timestamp, surrounded by + # quotation marks. Nanoseconds are ignored. + # - 'error_count' - Number of errors that have occurred while running. + # * The operator must be `=` or `!=` for status and inspected_storage. + # Examples: + # * inspected_storage = cloud_storage AND status = HEALTHY + # * inspected_storage = cloud_storage OR inspected_storage = bigquery + # * inspected_storage = cloud_storage AND (state = PAUSED OR state = HEALTHY) + # * last_run_time > \"2017-12-12T00:00:00+00:00\" + # The length of this field should be no more than 500 characters. # @param [String] order_by # Optional comma separated list of triggeredJob fields to order by, # followed by `asc` or `desc` postfix. This list is case-insensitive, @@ -1533,6 +1553,7 @@ module Google # Supported fields are: # - `create_time`: corresponds to time the JobTrigger was created. # - `update_time`: corresponds to time the JobTrigger was last updated. + # - `last_run_time`: corresponds to the last time the JobTrigger ran. # - `name`: corresponds to JobTrigger's name. # - `display_name`: corresponds to JobTrigger's display name. # - `status`: corresponds to JobTrigger's status. @@ -1559,11 +1580,12 @@ module Google # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification # @raise [Google::Apis::AuthorizationError] Authorization is required - def list_project_job_triggers(parent, order_by: nil, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block) + def list_project_job_triggers(parent, filter: nil, order_by: nil, page_size: nil, page_token: nil, fields: nil, quota_user: nil, options: nil, &block) command = make_simple_command(:get, 'v2/{+parent}/jobTriggers', options) command.response_representation = Google::Apis::DlpV2::GooglePrivacyDlpV2ListJobTriggersResponse::Representation command.response_class = Google::Apis::DlpV2::GooglePrivacyDlpV2ListJobTriggersResponse command.params['parent'] = parent unless parent.nil? + command.query['filter'] = filter unless filter.nil? command.query['orderBy'] = order_by unless order_by.nil? command.query['pageSize'] = page_size unless page_size.nil? command.query['pageToken'] = page_token unless page_token.nil? diff --git a/generated/google/apis/iam_v1.rb b/generated/google/apis/iam_v1.rb index d9a97cdf7..319db3fec 100644 --- a/generated/google/apis/iam_v1.rb +++ b/generated/google/apis/iam_v1.rb @@ -27,7 +27,7 @@ module Google # @see https://cloud.google.com/iam/ module IamV1 VERSION = 'V1' - REVISION = '20190111' + REVISION = '20190125' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/iam_v1/service.rb b/generated/google/apis/iam_v1/service.rb index d237b20b4..8637cffd4 100644 --- a/generated/google/apis/iam_v1/service.rb +++ b/generated/google/apis/iam_v1/service.rb @@ -871,6 +871,10 @@ module Google execute_or_queue_command(command, &block) end + # **Note**: This method is in the process of being deprecated. Call the + # [`signBlob()`](/iam/credentials/reference/rest/v1/projects.serviceAccounts/ + # signBlob) + # method of the Cloud IAM Service Account Credentials API instead. # Signs a blob using a service account's system-managed private key. # @param [String] name # The resource name of the service account in the following format: @@ -908,6 +912,10 @@ module Google execute_or_queue_command(command, &block) end + # **Note**: This method is in the process of being deprecated. Call the + # [`signJwt()`](/iam/credentials/reference/rest/v1/projects.serviceAccounts/ + # signJwt) + # method of the Cloud IAM Service Account Credentials API instead. # Signs a JWT using a service account's system-managed private key. # If no expiry time (`exp`) is provided in the `SignJwtRequest`, IAM sets an # an expiry time of one hour by default. If you request an expiry time of diff --git a/generated/google/apis/storagetransfer_v1.rb b/generated/google/apis/storagetransfer_v1.rb index 813c1c6b8..f0e2bd5ec 100644 --- a/generated/google/apis/storagetransfer_v1.rb +++ b/generated/google/apis/storagetransfer_v1.rb @@ -26,7 +26,7 @@ module Google # @see https://cloud.google.com/storage/transfer module StoragetransferV1 VERSION = 'V1' - REVISION = '20180919' + REVISION = '20190128' # View and manage your data across Google Cloud Platform services AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' diff --git a/generated/google/apis/storagetransfer_v1/classes.rb b/generated/google/apis/storagetransfer_v1/classes.rb index 4917e4352..b393f4610 100644 --- a/generated/google/apis/storagetransfer_v1/classes.rb +++ b/generated/google/apis/storagetransfer_v1/classes.rb @@ -187,6 +187,8 @@ module Google attr_accessor :error_count # Error samples. + # No more than 100 error log entries may be recorded for a given + # error code for a single task. # Corresponds to the JSON property `errorLogEntries` # @return [Array] attr_accessor :error_log_entries diff --git a/generated/google/apis/videointelligence_v1p2beta1.rb b/generated/google/apis/videointelligence_v1p2beta1.rb new file mode 100644 index 000000000..d9789a9a9 --- /dev/null +++ b/generated/google/apis/videointelligence_v1p2beta1.rb @@ -0,0 +1,35 @@ +# Copyright 2015 Google Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +require 'google/apis/videointelligence_v1p2beta1/service.rb' +require 'google/apis/videointelligence_v1p2beta1/classes.rb' +require 'google/apis/videointelligence_v1p2beta1/representations.rb' + +module Google + module Apis + # Cloud Video Intelligence API + # + # Detects objects, explicit content, and scene changes in videos. It also + # specifies the region for annotation and transcribes speech to text. + # + # @see https://cloud.google.com/video-intelligence/docs/ + module VideointelligenceV1p2beta1 + VERSION = 'V1p2beta1' + REVISION = '20190122' + + # View and manage your data across Google Cloud Platform services + AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform' + end + end +end diff --git a/generated/google/apis/videointelligence_v1p2beta1/classes.rb b/generated/google/apis/videointelligence_v1p2beta1/classes.rb new file mode 100644 index 000000000..938c5492f --- /dev/null +++ b/generated/google/apis/videointelligence_v1p2beta1/classes.rb @@ -0,0 +1,2872 @@ +# Copyright 2015 Google Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +require 'date' +require 'google/apis/core/base_service' +require 'google/apis/core/json_representation' +require 'google/apis/core/hashable' +require 'google/apis/errors' + +module Google + module Apis + module VideointelligenceV1p2beta1 + + # Video annotation progress. Included in the `metadata` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1AnnotateVideoProgress + include Google::Apis::Core::Hashable + + # Progress metadata for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationProgress` + # @return [Array] + attr_accessor :annotation_progress + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_progress = args[:annotation_progress] if args.key?(:annotation_progress) + end + end + + # Video annotation response. Included in the `response` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1AnnotateVideoResponse + include Google::Apis::Core::Hashable + + # Annotation results for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationResults` + # @return [Array] + attr_accessor :annotation_results + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_results = args[:annotation_results] if args.key?(:annotation_results) + end + end + + # Detected entity from video analysis. + class GoogleCloudVideointelligenceV1Entity + include Google::Apis::Core::Hashable + + # Textual description, e.g. `Fixed-gear bicycle`. + # Corresponds to the JSON property `description` + # @return [String] + attr_accessor :description + + # Opaque entity ID. Some IDs may be available in + # [Google Knowledge Graph Search + # API](https://developers.google.com/knowledge-graph/). + # Corresponds to the JSON property `entityId` + # @return [String] + attr_accessor :entity_id + + # Language code for `description` in BCP-47 format. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @description = args[:description] if args.key?(:description) + @entity_id = args[:entity_id] if args.key?(:entity_id) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + class GoogleCloudVideointelligenceV1ExplicitContentAnnotation + include Google::Apis::Core::Hashable + + # All video frames where explicit content was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @frames = args[:frames] if args.key?(:frames) + end + end + + # Video frame level annotation results for explicit content. + class GoogleCloudVideointelligenceV1ExplicitContentFrame + include Google::Apis::Core::Hashable + + # Likelihood of the pornography content.. + # Corresponds to the JSON property `pornographyLikelihood` + # @return [String] + attr_accessor :pornography_likelihood + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Label annotation. + class GoogleCloudVideointelligenceV1LabelAnnotation + include Google::Apis::Core::Hashable + + # Common categories for the detected entity. + # E.g. when the label is `Terrier` the category is likely `dog`. And in some + # cases there might be more than one categories e.g. `Terrier` could also be + # a `pet`. + # Corresponds to the JSON property `categoryEntities` + # @return [Array] + attr_accessor :category_entities + + # Detected entity from video analysis. + # Corresponds to the JSON property `entity` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1Entity] + attr_accessor :entity + + # All video frames where a label was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # All video segments where a label was detected. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @category_entities = args[:category_entities] if args.key?(:category_entities) + @entity = args[:entity] if args.key?(:entity) + @frames = args[:frames] if args.key?(:frames) + @segments = args[:segments] if args.key?(:segments) + end + end + + # Video frame level annotation results for label detection. + class GoogleCloudVideointelligenceV1LabelFrame + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Video segment level annotation results for label detection. + class GoogleCloudVideointelligenceV1LabelSegment + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoSegment] + attr_accessor :segment + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @segment = args[:segment] if args.key?(:segment) + end + end + + # Alternative hypotheses (a.k.a. n-best list). + class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative + include Google::Apis::Core::Hashable + + # The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is typically provided only for the top hypothesis, and + # only for `is_final=true` results. Clients should not rely on the + # `confidence` field as it is not guaranteed to be accurate or consistent. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Transcript text representing the words that the user spoke. + # Corresponds to the JSON property `transcript` + # @return [String] + attr_accessor :transcript + + # A list of word-specific information for each recognized word. + # Corresponds to the JSON property `words` + # @return [Array] + attr_accessor :words + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @transcript = args[:transcript] if args.key?(:transcript) + @words = args[:words] if args.key?(:words) + end + end + + # A speech recognition result corresponding to a portion of the audio. + class GoogleCloudVideointelligenceV1SpeechTranscription + include Google::Apis::Core::Hashable + + # May contain one or more recognition hypotheses (up to the maximum specified + # in `max_alternatives`). These alternatives are ordered in terms of + # accuracy, with the top (first) alternative being the most probable, as + # ranked by the recognizer. + # Corresponds to the JSON property `alternatives` + # @return [Array] + attr_accessor :alternatives + + # Output only. The + # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the + # language in this result. This language code was detected to have the most + # likelihood of being spoken in the audio. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @alternatives = args[:alternatives] if args.key?(:alternatives) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Annotation progress for a single video. + class GoogleCloudVideointelligenceV1VideoAnnotationProgress + include Google::Apis::Core::Hashable + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Approximate percentage processed thus far. Guaranteed to be + # 100 when fully processed. + # Corresponds to the JSON property `progressPercent` + # @return [Fixnum] + attr_accessor :progress_percent + + # Time when the request was received. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # Time of the most recent update. + # Corresponds to the JSON property `updateTime` + # @return [String] + attr_accessor :update_time + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @progress_percent = args[:progress_percent] if args.key?(:progress_percent) + @start_time = args[:start_time] if args.key?(:start_time) + @update_time = args[:update_time] if args.key?(:update_time) + end + end + + # Annotation results for a single video. + class GoogleCloudVideointelligenceV1VideoAnnotationResults + include Google::Apis::Core::Hashable + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + # Corresponds to the JSON property `error` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus] + attr_accessor :error + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + # Corresponds to the JSON property `explicitAnnotation` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1ExplicitContentAnnotation] + attr_accessor :explicit_annotation + + # Label annotations on frame level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `frameLabelAnnotations` + # @return [Array] + attr_accessor :frame_label_annotations + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Label annotations on video level or user specified segment level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `segmentLabelAnnotations` + # @return [Array] + attr_accessor :segment_label_annotations + + # Shot annotations. Each shot is represented as a video segment. + # Corresponds to the JSON property `shotAnnotations` + # @return [Array] + attr_accessor :shot_annotations + + # Label annotations on shot level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `shotLabelAnnotations` + # @return [Array] + attr_accessor :shot_label_annotations + + # Speech transcription. + # Corresponds to the JSON property `speechTranscriptions` + # @return [Array] + attr_accessor :speech_transcriptions + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @error = args[:error] if args.key?(:error) + @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation) + @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations) + @shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations) + @shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations) + @speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions) + end + end + + # Video segment. + class GoogleCloudVideointelligenceV1VideoSegment + include Google::Apis::Core::Hashable + + # Time-offset, relative to the beginning of the video, + # corresponding to the end of the segment (inclusive). + # Corresponds to the JSON property `endTimeOffset` + # @return [String] + attr_accessor :end_time_offset + + # Time-offset, relative to the beginning of the video, + # corresponding to the start of the segment (inclusive). + # Corresponds to the JSON property `startTimeOffset` + # @return [String] + attr_accessor :start_time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @end_time_offset = args[:end_time_offset] if args.key?(:end_time_offset) + @start_time_offset = args[:start_time_offset] if args.key?(:start_time_offset) + end + end + + # Word-specific information for recognized words. Word information is only + # included in the response when certain request parameters are set, such + # as `enable_word_time_offsets`. + class GoogleCloudVideointelligenceV1WordInfo + include Google::Apis::Core::Hashable + + # Output only. The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is set only for the top alternative. + # This field is not guaranteed to be accurate and users should not rely on it + # to be always provided. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time offset relative to the beginning of the audio, and + # corresponding to the end of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `endTime` + # @return [String] + attr_accessor :end_time + + # Output only. A distinct integer value is assigned for every speaker within + # the audio. This field specifies which one of those speakers was detected to + # have spoken this word. Value ranges from 1 up to diarization_speaker_count, + # and is only set if speaker diarization is enabled. + # Corresponds to the JSON property `speakerTag` + # @return [Fixnum] + attr_accessor :speaker_tag + + # Time offset relative to the beginning of the audio, and + # corresponding to the start of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # The word corresponding to this set of information. + # Corresponds to the JSON property `word` + # @return [String] + attr_accessor :word + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @end_time = args[:end_time] if args.key?(:end_time) + @speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag) + @start_time = args[:start_time] if args.key?(:start_time) + @word = args[:word] if args.key?(:word) + end + end + + # Video annotation progress. Included in the `metadata` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress + include Google::Apis::Core::Hashable + + # Progress metadata for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationProgress` + # @return [Array] + attr_accessor :annotation_progress + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_progress = args[:annotation_progress] if args.key?(:annotation_progress) + end + end + + # Video annotation response. Included in the `response` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse + include Google::Apis::Core::Hashable + + # Annotation results for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationResults` + # @return [Array] + attr_accessor :annotation_results + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_results = args[:annotation_results] if args.key?(:annotation_results) + end + end + + # Detected entity from video analysis. + class GoogleCloudVideointelligenceV1beta2Entity + include Google::Apis::Core::Hashable + + # Textual description, e.g. `Fixed-gear bicycle`. + # Corresponds to the JSON property `description` + # @return [String] + attr_accessor :description + + # Opaque entity ID. Some IDs may be available in + # [Google Knowledge Graph Search + # API](https://developers.google.com/knowledge-graph/). + # Corresponds to the JSON property `entityId` + # @return [String] + attr_accessor :entity_id + + # Language code for `description` in BCP-47 format. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @description = args[:description] if args.key?(:description) + @entity_id = args[:entity_id] if args.key?(:entity_id) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + class GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation + include Google::Apis::Core::Hashable + + # All video frames where explicit content was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @frames = args[:frames] if args.key?(:frames) + end + end + + # Video frame level annotation results for explicit content. + class GoogleCloudVideointelligenceV1beta2ExplicitContentFrame + include Google::Apis::Core::Hashable + + # Likelihood of the pornography content.. + # Corresponds to the JSON property `pornographyLikelihood` + # @return [String] + attr_accessor :pornography_likelihood + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Label annotation. + class GoogleCloudVideointelligenceV1beta2LabelAnnotation + include Google::Apis::Core::Hashable + + # Common categories for the detected entity. + # E.g. when the label is `Terrier` the category is likely `dog`. And in some + # cases there might be more than one categories e.g. `Terrier` could also be + # a `pet`. + # Corresponds to the JSON property `categoryEntities` + # @return [Array] + attr_accessor :category_entities + + # Detected entity from video analysis. + # Corresponds to the JSON property `entity` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2Entity] + attr_accessor :entity + + # All video frames where a label was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # All video segments where a label was detected. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @category_entities = args[:category_entities] if args.key?(:category_entities) + @entity = args[:entity] if args.key?(:entity) + @frames = args[:frames] if args.key?(:frames) + @segments = args[:segments] if args.key?(:segments) + end + end + + # Video frame level annotation results for label detection. + class GoogleCloudVideointelligenceV1beta2LabelFrame + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Video segment level annotation results for label detection. + class GoogleCloudVideointelligenceV1beta2LabelSegment + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoSegment] + attr_accessor :segment + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @segment = args[:segment] if args.key?(:segment) + end + end + + # Alternative hypotheses (a.k.a. n-best list). + class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative + include Google::Apis::Core::Hashable + + # The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is typically provided only for the top hypothesis, and + # only for `is_final=true` results. Clients should not rely on the + # `confidence` field as it is not guaranteed to be accurate or consistent. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Transcript text representing the words that the user spoke. + # Corresponds to the JSON property `transcript` + # @return [String] + attr_accessor :transcript + + # A list of word-specific information for each recognized word. + # Corresponds to the JSON property `words` + # @return [Array] + attr_accessor :words + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @transcript = args[:transcript] if args.key?(:transcript) + @words = args[:words] if args.key?(:words) + end + end + + # A speech recognition result corresponding to a portion of the audio. + class GoogleCloudVideointelligenceV1beta2SpeechTranscription + include Google::Apis::Core::Hashable + + # May contain one or more recognition hypotheses (up to the maximum specified + # in `max_alternatives`). These alternatives are ordered in terms of + # accuracy, with the top (first) alternative being the most probable, as + # ranked by the recognizer. + # Corresponds to the JSON property `alternatives` + # @return [Array] + attr_accessor :alternatives + + # Output only. The + # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the + # language in this result. This language code was detected to have the most + # likelihood of being spoken in the audio. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @alternatives = args[:alternatives] if args.key?(:alternatives) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Annotation progress for a single video. + class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress + include Google::Apis::Core::Hashable + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Approximate percentage processed thus far. Guaranteed to be + # 100 when fully processed. + # Corresponds to the JSON property `progressPercent` + # @return [Fixnum] + attr_accessor :progress_percent + + # Time when the request was received. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # Time of the most recent update. + # Corresponds to the JSON property `updateTime` + # @return [String] + attr_accessor :update_time + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @progress_percent = args[:progress_percent] if args.key?(:progress_percent) + @start_time = args[:start_time] if args.key?(:start_time) + @update_time = args[:update_time] if args.key?(:update_time) + end + end + + # Annotation results for a single video. + class GoogleCloudVideointelligenceV1beta2VideoAnnotationResults + include Google::Apis::Core::Hashable + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + # Corresponds to the JSON property `error` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus] + attr_accessor :error + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + # Corresponds to the JSON property `explicitAnnotation` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation] + attr_accessor :explicit_annotation + + # Label annotations on frame level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `frameLabelAnnotations` + # @return [Array] + attr_accessor :frame_label_annotations + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Label annotations on video level or user specified segment level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `segmentLabelAnnotations` + # @return [Array] + attr_accessor :segment_label_annotations + + # Shot annotations. Each shot is represented as a video segment. + # Corresponds to the JSON property `shotAnnotations` + # @return [Array] + attr_accessor :shot_annotations + + # Label annotations on shot level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `shotLabelAnnotations` + # @return [Array] + attr_accessor :shot_label_annotations + + # Speech transcription. + # Corresponds to the JSON property `speechTranscriptions` + # @return [Array] + attr_accessor :speech_transcriptions + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @error = args[:error] if args.key?(:error) + @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation) + @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations) + @shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations) + @shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations) + @speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions) + end + end + + # Video segment. + class GoogleCloudVideointelligenceV1beta2VideoSegment + include Google::Apis::Core::Hashable + + # Time-offset, relative to the beginning of the video, + # corresponding to the end of the segment (inclusive). + # Corresponds to the JSON property `endTimeOffset` + # @return [String] + attr_accessor :end_time_offset + + # Time-offset, relative to the beginning of the video, + # corresponding to the start of the segment (inclusive). + # Corresponds to the JSON property `startTimeOffset` + # @return [String] + attr_accessor :start_time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @end_time_offset = args[:end_time_offset] if args.key?(:end_time_offset) + @start_time_offset = args[:start_time_offset] if args.key?(:start_time_offset) + end + end + + # Word-specific information for recognized words. Word information is only + # included in the response when certain request parameters are set, such + # as `enable_word_time_offsets`. + class GoogleCloudVideointelligenceV1beta2WordInfo + include Google::Apis::Core::Hashable + + # Output only. The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is set only for the top alternative. + # This field is not guaranteed to be accurate and users should not rely on it + # to be always provided. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time offset relative to the beginning of the audio, and + # corresponding to the end of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `endTime` + # @return [String] + attr_accessor :end_time + + # Output only. A distinct integer value is assigned for every speaker within + # the audio. This field specifies which one of those speakers was detected to + # have spoken this word. Value ranges from 1 up to diarization_speaker_count, + # and is only set if speaker diarization is enabled. + # Corresponds to the JSON property `speakerTag` + # @return [Fixnum] + attr_accessor :speaker_tag + + # Time offset relative to the beginning of the audio, and + # corresponding to the start of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # The word corresponding to this set of information. + # Corresponds to the JSON property `word` + # @return [String] + attr_accessor :word + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @end_time = args[:end_time] if args.key?(:end_time) + @speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag) + @start_time = args[:start_time] if args.key?(:start_time) + @word = args[:word] if args.key?(:word) + end + end + + # Video annotation progress. Included in the `metadata` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress + include Google::Apis::Core::Hashable + + # Progress metadata for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationProgress` + # @return [Array] + attr_accessor :annotation_progress + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_progress = args[:annotation_progress] if args.key?(:annotation_progress) + end + end + + # Video annotation response. Included in the `response` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoResponse + include Google::Apis::Core::Hashable + + # Annotation results for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationResults` + # @return [Array] + attr_accessor :annotation_results + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_results = args[:annotation_results] if args.key?(:annotation_results) + end + end + + # Detected entity from video analysis. + class GoogleCloudVideointelligenceV1p1beta1Entity + include Google::Apis::Core::Hashable + + # Textual description, e.g. `Fixed-gear bicycle`. + # Corresponds to the JSON property `description` + # @return [String] + attr_accessor :description + + # Opaque entity ID. Some IDs may be available in + # [Google Knowledge Graph Search + # API](https://developers.google.com/knowledge-graph/). + # Corresponds to the JSON property `entityId` + # @return [String] + attr_accessor :entity_id + + # Language code for `description` in BCP-47 format. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @description = args[:description] if args.key?(:description) + @entity_id = args[:entity_id] if args.key?(:entity_id) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation + include Google::Apis::Core::Hashable + + # All video frames where explicit content was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @frames = args[:frames] if args.key?(:frames) + end + end + + # Video frame level annotation results for explicit content. + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame + include Google::Apis::Core::Hashable + + # Likelihood of the pornography content.. + # Corresponds to the JSON property `pornographyLikelihood` + # @return [String] + attr_accessor :pornography_likelihood + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Label annotation. + class GoogleCloudVideointelligenceV1p1beta1LabelAnnotation + include Google::Apis::Core::Hashable + + # Common categories for the detected entity. + # E.g. when the label is `Terrier` the category is likely `dog`. And in some + # cases there might be more than one categories e.g. `Terrier` could also be + # a `pet`. + # Corresponds to the JSON property `categoryEntities` + # @return [Array] + attr_accessor :category_entities + + # Detected entity from video analysis. + # Corresponds to the JSON property `entity` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1Entity] + attr_accessor :entity + + # All video frames where a label was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # All video segments where a label was detected. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @category_entities = args[:category_entities] if args.key?(:category_entities) + @entity = args[:entity] if args.key?(:entity) + @frames = args[:frames] if args.key?(:frames) + @segments = args[:segments] if args.key?(:segments) + end + end + + # Video frame level annotation results for label detection. + class GoogleCloudVideointelligenceV1p1beta1LabelFrame + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Video segment level annotation results for label detection. + class GoogleCloudVideointelligenceV1p1beta1LabelSegment + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment] + attr_accessor :segment + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @segment = args[:segment] if args.key?(:segment) + end + end + + # Alternative hypotheses (a.k.a. n-best list). + class GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative + include Google::Apis::Core::Hashable + + # The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is typically provided only for the top hypothesis, and + # only for `is_final=true` results. Clients should not rely on the + # `confidence` field as it is not guaranteed to be accurate or consistent. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Transcript text representing the words that the user spoke. + # Corresponds to the JSON property `transcript` + # @return [String] + attr_accessor :transcript + + # A list of word-specific information for each recognized word. + # Corresponds to the JSON property `words` + # @return [Array] + attr_accessor :words + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @transcript = args[:transcript] if args.key?(:transcript) + @words = args[:words] if args.key?(:words) + end + end + + # A speech recognition result corresponding to a portion of the audio. + class GoogleCloudVideointelligenceV1p1beta1SpeechTranscription + include Google::Apis::Core::Hashable + + # May contain one or more recognition hypotheses (up to the maximum specified + # in `max_alternatives`). These alternatives are ordered in terms of + # accuracy, with the top (first) alternative being the most probable, as + # ranked by the recognizer. + # Corresponds to the JSON property `alternatives` + # @return [Array] + attr_accessor :alternatives + + # Output only. The + # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the + # language in this result. This language code was detected to have the most + # likelihood of being spoken in the audio. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @alternatives = args[:alternatives] if args.key?(:alternatives) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Annotation progress for a single video. + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress + include Google::Apis::Core::Hashable + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Approximate percentage processed thus far. Guaranteed to be + # 100 when fully processed. + # Corresponds to the JSON property `progressPercent` + # @return [Fixnum] + attr_accessor :progress_percent + + # Time when the request was received. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # Time of the most recent update. + # Corresponds to the JSON property `updateTime` + # @return [String] + attr_accessor :update_time + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @progress_percent = args[:progress_percent] if args.key?(:progress_percent) + @start_time = args[:start_time] if args.key?(:start_time) + @update_time = args[:update_time] if args.key?(:update_time) + end + end + + # Annotation results for a single video. + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults + include Google::Apis::Core::Hashable + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + # Corresponds to the JSON property `error` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus] + attr_accessor :error + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + # Corresponds to the JSON property `explicitAnnotation` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation] + attr_accessor :explicit_annotation + + # Label annotations on frame level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `frameLabelAnnotations` + # @return [Array] + attr_accessor :frame_label_annotations + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Label annotations on video level or user specified segment level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `segmentLabelAnnotations` + # @return [Array] + attr_accessor :segment_label_annotations + + # Shot annotations. Each shot is represented as a video segment. + # Corresponds to the JSON property `shotAnnotations` + # @return [Array] + attr_accessor :shot_annotations + + # Label annotations on shot level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `shotLabelAnnotations` + # @return [Array] + attr_accessor :shot_label_annotations + + # Speech transcription. + # Corresponds to the JSON property `speechTranscriptions` + # @return [Array] + attr_accessor :speech_transcriptions + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @error = args[:error] if args.key?(:error) + @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation) + @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations) + @shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations) + @shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations) + @speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions) + end + end + + # Video segment. + class GoogleCloudVideointelligenceV1p1beta1VideoSegment + include Google::Apis::Core::Hashable + + # Time-offset, relative to the beginning of the video, + # corresponding to the end of the segment (inclusive). + # Corresponds to the JSON property `endTimeOffset` + # @return [String] + attr_accessor :end_time_offset + + # Time-offset, relative to the beginning of the video, + # corresponding to the start of the segment (inclusive). + # Corresponds to the JSON property `startTimeOffset` + # @return [String] + attr_accessor :start_time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @end_time_offset = args[:end_time_offset] if args.key?(:end_time_offset) + @start_time_offset = args[:start_time_offset] if args.key?(:start_time_offset) + end + end + + # Word-specific information for recognized words. Word information is only + # included in the response when certain request parameters are set, such + # as `enable_word_time_offsets`. + class GoogleCloudVideointelligenceV1p1beta1WordInfo + include Google::Apis::Core::Hashable + + # Output only. The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is set only for the top alternative. + # This field is not guaranteed to be accurate and users should not rely on it + # to be always provided. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time offset relative to the beginning of the audio, and + # corresponding to the end of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `endTime` + # @return [String] + attr_accessor :end_time + + # Output only. A distinct integer value is assigned for every speaker within + # the audio. This field specifies which one of those speakers was detected to + # have spoken this word. Value ranges from 1 up to diarization_speaker_count, + # and is only set if speaker diarization is enabled. + # Corresponds to the JSON property `speakerTag` + # @return [Fixnum] + attr_accessor :speaker_tag + + # Time offset relative to the beginning of the audio, and + # corresponding to the start of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # The word corresponding to this set of information. + # Corresponds to the JSON property `word` + # @return [String] + attr_accessor :word + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @end_time = args[:end_time] if args.key?(:end_time) + @speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag) + @start_time = args[:start_time] if args.key?(:start_time) + @word = args[:word] if args.key?(:word) + end + end + + # Video annotation progress. Included in the `metadata` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress + include Google::Apis::Core::Hashable + + # Progress metadata for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationProgress` + # @return [Array] + attr_accessor :annotation_progress + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_progress = args[:annotation_progress] if args.key?(:annotation_progress) + end + end + + # Video annotation request. + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoRequest + include Google::Apis::Core::Hashable + + # Requested video annotation features. + # Corresponds to the JSON property `features` + # @return [Array] + attr_accessor :features + + # The video data bytes. + # If unset, the input video(s) should be specified via `input_uri`. + # If set, `input_uri` should be unset. + # Corresponds to the JSON property `inputContent` + # NOTE: Values are automatically base64 encoded/decoded in the client library. + # @return [String] + attr_accessor :input_content + + # Input video location. Currently, only + # [Google Cloud Storage](https://cloud.google.com/storage/) URIs are + # supported, which must be specified in the following format: + # `gs://bucket-id/object-id` (other URI formats return + # google.rpc.Code.INVALID_ARGUMENT). For more information, see + # [Request URIs](/storage/docs/reference-uris). + # A video URI may include wildcards in `object-id`, and thus identify + # multiple videos. Supported wildcards: '*' to match 0 or more characters; + # '?' to match 1 character. If unset, the input video should be embedded + # in the request as `input_content`. If set, `input_content` should be unset. + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Optional cloud region where annotation should take place. Supported cloud + # regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region + # is specified, a region will be determined based on video file location. + # Corresponds to the JSON property `locationId` + # @return [String] + attr_accessor :location_id + + # Optional location where the output (in JSON format) should be stored. + # Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) + # URIs are supported, which must be specified in the following format: + # `gs://bucket-id/object-id` (other URI formats return + # google.rpc.Code.INVALID_ARGUMENT). For more information, see + # [Request URIs](/storage/docs/reference-uris). + # Corresponds to the JSON property `outputUri` + # @return [String] + attr_accessor :output_uri + + # Video context and/or feature-specific parameters. + # Corresponds to the JSON property `videoContext` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoContext] + attr_accessor :video_context + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @features = args[:features] if args.key?(:features) + @input_content = args[:input_content] if args.key?(:input_content) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @location_id = args[:location_id] if args.key?(:location_id) + @output_uri = args[:output_uri] if args.key?(:output_uri) + @video_context = args[:video_context] if args.key?(:video_context) + end + end + + # Video annotation response. Included in the `response` + # field of the `Operation` returned by the `GetOperation` + # call of the `google::longrunning::Operations` service. + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse + include Google::Apis::Core::Hashable + + # Annotation results for all videos specified in `AnnotateVideoRequest`. + # Corresponds to the JSON property `annotationResults` + # @return [Array] + attr_accessor :annotation_results + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @annotation_results = args[:annotation_results] if args.key?(:annotation_results) + end + end + + # Detected entity from video analysis. + class GoogleCloudVideointelligenceV1p2beta1Entity + include Google::Apis::Core::Hashable + + # Textual description, e.g. `Fixed-gear bicycle`. + # Corresponds to the JSON property `description` + # @return [String] + attr_accessor :description + + # Opaque entity ID. Some IDs may be available in + # [Google Knowledge Graph Search + # API](https://developers.google.com/knowledge-graph/). + # Corresponds to the JSON property `entityId` + # @return [String] + attr_accessor :entity_id + + # Language code for `description` in BCP-47 format. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @description = args[:description] if args.key?(:description) + @entity_id = args[:entity_id] if args.key?(:entity_id) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation + include Google::Apis::Core::Hashable + + # All video frames where explicit content was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @frames = args[:frames] if args.key?(:frames) + end + end + + # Config for EXPLICIT_CONTENT_DETECTION. + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig + include Google::Apis::Core::Hashable + + # Model to use for explicit content detection. + # Supported values: "builtin/stable" (the default if unset) and + # "builtin/latest". + # Corresponds to the JSON property `model` + # @return [String] + attr_accessor :model + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @model = args[:model] if args.key?(:model) + end + end + + # Video frame level annotation results for explicit content. + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame + include Google::Apis::Core::Hashable + + # Likelihood of the pornography content.. + # Corresponds to the JSON property `pornographyLikelihood` + # @return [String] + attr_accessor :pornography_likelihood + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Label annotation. + class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation + include Google::Apis::Core::Hashable + + # Common categories for the detected entity. + # E.g. when the label is `Terrier` the category is likely `dog`. And in some + # cases there might be more than one categories e.g. `Terrier` could also be + # a `pet`. + # Corresponds to the JSON property `categoryEntities` + # @return [Array] + attr_accessor :category_entities + + # Detected entity from video analysis. + # Corresponds to the JSON property `entity` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity] + attr_accessor :entity + + # All video frames where a label was detected. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # All video segments where a label was detected. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @category_entities = args[:category_entities] if args.key?(:category_entities) + @entity = args[:entity] if args.key?(:entity) + @frames = args[:frames] if args.key?(:frames) + @segments = args[:segments] if args.key?(:segments) + end + end + + # Config for LABEL_DETECTION. + class GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig + include Google::Apis::Core::Hashable + + # What labels should be detected with LABEL_DETECTION, in addition to + # video-level labels or segment-level labels. + # If unspecified, defaults to `SHOT_MODE`. + # Corresponds to the JSON property `labelDetectionMode` + # @return [String] + attr_accessor :label_detection_mode + + # Model to use for label detection. + # Supported values: "builtin/stable" (the default if unset) and + # "builtin/latest". + # Corresponds to the JSON property `model` + # @return [String] + attr_accessor :model + + # Whether the video has been shot from a stationary (i.e. non-moving) camera. + # When set to true, might improve detection accuracy for moving objects. + # Should be used with `SHOT_AND_FRAME_MODE` enabled. + # Corresponds to the JSON property `stationaryCamera` + # @return [Boolean] + attr_accessor :stationary_camera + alias_method :stationary_camera?, :stationary_camera + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @label_detection_mode = args[:label_detection_mode] if args.key?(:label_detection_mode) + @model = args[:model] if args.key?(:model) + @stationary_camera = args[:stationary_camera] if args.key?(:stationary_camera) + end + end + + # Video frame level annotation results for label detection. + class GoogleCloudVideointelligenceV1p2beta1LabelFrame + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time-offset, relative to the beginning of the video, corresponding to the + # video frame for this location. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Video segment level annotation results for label detection. + class GoogleCloudVideointelligenceV1p2beta1LabelSegment + include Google::Apis::Core::Hashable + + # Confidence that the label is accurate. Range: [0, 1]. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment] + attr_accessor :segment + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @segment = args[:segment] if args.key?(:segment) + end + end + + # Normalized bounding box. + # The normalized vertex coordinates are relative to the original image. + # Range: [0, 1]. + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox + include Google::Apis::Core::Hashable + + # Bottom Y coordinate. + # Corresponds to the JSON property `bottom` + # @return [Float] + attr_accessor :bottom + + # Left X coordinate. + # Corresponds to the JSON property `left` + # @return [Float] + attr_accessor :left + + # Right X coordinate. + # Corresponds to the JSON property `right` + # @return [Float] + attr_accessor :right + + # Top Y coordinate. + # Corresponds to the JSON property `top` + # @return [Float] + attr_accessor :top + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @bottom = args[:bottom] if args.key?(:bottom) + @left = args[:left] if args.key?(:left) + @right = args[:right] if args.key?(:right) + @top = args[:top] if args.key?(:top) + end + end + + # Normalized bounding polygon for text (that might not be aligned with axis). + # Contains list of the corner points in clockwise order starting from + # top-left corner. For example, for a rectangular bounding box: + # When the text is horizontal it might look like: + # 0----1 + # | | + # 3----2 + # When it's clockwise rotated 180 degrees around the top-left corner it + # becomes: + # 2----3 + # | | + # 1----0 + # and the vertex order will still be (0, 1, 2, 3). Note that values can be less + # than 0, or greater than 1 due to trignometric calculations for location of + # the box. + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly + include Google::Apis::Core::Hashable + + # Normalized vertices of the bounding polygon. + # Corresponds to the JSON property `vertices` + # @return [Array] + attr_accessor :vertices + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @vertices = args[:vertices] if args.key?(:vertices) + end + end + + # A vertex represents a 2D point in the image. + # NOTE: the normalized vertex coordinates are relative to the original image + # and range from 0 to 1. + class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex + include Google::Apis::Core::Hashable + + # X coordinate. + # Corresponds to the JSON property `x` + # @return [Float] + attr_accessor :x + + # Y coordinate. + # Corresponds to the JSON property `y` + # @return [Float] + attr_accessor :y + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @x = args[:x] if args.key?(:x) + @y = args[:y] if args.key?(:y) + end + end + + # Annotations corresponding to one tracked object. + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation + include Google::Apis::Core::Hashable + + # Object category's labeling confidence of this track. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Detected entity from video analysis. + # Corresponds to the JSON property `entity` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity] + attr_accessor :entity + + # Information corresponding to all frames where this object track appears. + # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame + # messages in frames. + # Streaming mode: it can only be one ObjectTrackingFrame message in frames. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment] + attr_accessor :segment + + # Streaming mode ONLY. + # In streaming mode, we do not know the end time of a tracked object + # before it is completed. Hence, there is no VideoSegment info returned. + # Instead, we provide a unique identifiable integer track_id so that + # the customers can correlate the results of the ongoing + # ObjectTrackAnnotation of the same track_id over time. + # Corresponds to the JSON property `trackId` + # @return [Fixnum] + attr_accessor :track_id + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @entity = args[:entity] if args.key?(:entity) + @frames = args[:frames] if args.key?(:frames) + @segment = args[:segment] if args.key?(:segment) + @track_id = args[:track_id] if args.key?(:track_id) + end + end + + # Video frame level annotations for object detection and tracking. This field + # stores per frame location, time offset, and confidence. + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame + include Google::Apis::Core::Hashable + + # Normalized bounding box. + # The normalized vertex coordinates are relative to the original image. + # Range: [0, 1]. + # Corresponds to the JSON property `normalizedBoundingBox` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox] + attr_accessor :normalized_bounding_box + + # The timestamp of the frame in microseconds. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @normalized_bounding_box = args[:normalized_bounding_box] if args.key?(:normalized_bounding_box) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Config for SHOT_CHANGE_DETECTION. + class GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig + include Google::Apis::Core::Hashable + + # Model to use for shot change detection. + # Supported values: "builtin/stable" (the default if unset) and + # "builtin/latest". + # Corresponds to the JSON property `model` + # @return [String] + attr_accessor :model + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @model = args[:model] if args.key?(:model) + end + end + + # Provides "hints" to the speech recognizer to favor specific words and phrases + # in the results. + class GoogleCloudVideointelligenceV1p2beta1SpeechContext + include Google::Apis::Core::Hashable + + # *Optional* A list of strings containing words and phrases "hints" so that + # the speech recognition is more likely to recognize them. This can be used + # to improve the accuracy for specific words and phrases, for example, if + # specific commands are typically spoken by the user. This can also be used + # to add additional words to the vocabulary of the recognizer. See + # [usage limits](https://cloud.google.com/speech/limits#content). + # Corresponds to the JSON property `phrases` + # @return [Array] + attr_accessor :phrases + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @phrases = args[:phrases] if args.key?(:phrases) + end + end + + # Alternative hypotheses (a.k.a. n-best list). + class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative + include Google::Apis::Core::Hashable + + # The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is typically provided only for the top hypothesis, and + # only for `is_final=true` results. Clients should not rely on the + # `confidence` field as it is not guaranteed to be accurate or consistent. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Transcript text representing the words that the user spoke. + # Corresponds to the JSON property `transcript` + # @return [String] + attr_accessor :transcript + + # A list of word-specific information for each recognized word. + # Corresponds to the JSON property `words` + # @return [Array] + attr_accessor :words + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @transcript = args[:transcript] if args.key?(:transcript) + @words = args[:words] if args.key?(:words) + end + end + + # A speech recognition result corresponding to a portion of the audio. + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription + include Google::Apis::Core::Hashable + + # May contain one or more recognition hypotheses (up to the maximum specified + # in `max_alternatives`). These alternatives are ordered in terms of + # accuracy, with the top (first) alternative being the most probable, as + # ranked by the recognizer. + # Corresponds to the JSON property `alternatives` + # @return [Array] + attr_accessor :alternatives + + # Output only. The + # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag of the + # language in this result. This language code was detected to have the most + # likelihood of being spoken in the audio. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @alternatives = args[:alternatives] if args.key?(:alternatives) + @language_code = args[:language_code] if args.key?(:language_code) + end + end + + # Config for SPEECH_TRANSCRIPTION. + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig + include Google::Apis::Core::Hashable + + # *Optional* For file formats, such as MXF or MKV, supporting multiple audio + # tracks, specify up to two tracks. Default: track 0. + # Corresponds to the JSON property `audioTracks` + # @return [Array] + attr_accessor :audio_tracks + + # *Optional* + # If set, specifies the estimated number of speakers in the conversation. + # If not set, defaults to '2'. + # Ignored unless enable_speaker_diarization is set to true. + # Corresponds to the JSON property `diarizationSpeakerCount` + # @return [Fixnum] + attr_accessor :diarization_speaker_count + + # *Optional* If 'true', adds punctuation to recognition result hypotheses. + # This feature is only available in select languages. Setting this for + # requests in other languages has no effect at all. The default 'false' value + # does not add punctuation to result hypotheses. NOTE: "This is currently + # offered as an experimental service, complimentary to all users. In the + # future this may be exclusively available as a premium feature." + # Corresponds to the JSON property `enableAutomaticPunctuation` + # @return [Boolean] + attr_accessor :enable_automatic_punctuation + alias_method :enable_automatic_punctuation?, :enable_automatic_punctuation + + # *Optional* If 'true', enables speaker detection for each recognized word in + # the top alternative of the recognition result using a speaker_tag provided + # in the WordInfo. + # Note: When this is true, we send all the words from the beginning of the + # audio for the top alternative in every consecutive responses. + # This is done in order to improve our speaker tags as our models learn to + # identify the speakers in the conversation over time. + # Corresponds to the JSON property `enableSpeakerDiarization` + # @return [Boolean] + attr_accessor :enable_speaker_diarization + alias_method :enable_speaker_diarization?, :enable_speaker_diarization + + # *Optional* If `true`, the top result includes a list of words and the + # confidence for those words. If `false`, no word-level confidence + # information is returned. The default is `false`. + # Corresponds to the JSON property `enableWordConfidence` + # @return [Boolean] + attr_accessor :enable_word_confidence + alias_method :enable_word_confidence?, :enable_word_confidence + + # *Optional* If set to `true`, the server will attempt to filter out + # profanities, replacing all but the initial character in each filtered word + # with asterisks, e.g. "f***". If set to `false` or omitted, profanities + # won't be filtered out. + # Corresponds to the JSON property `filterProfanity` + # @return [Boolean] + attr_accessor :filter_profanity + alias_method :filter_profanity?, :filter_profanity + + # *Required* The language of the supplied audio as a + # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. + # Example: "en-US". + # See [Language Support](https://cloud.google.com/speech/docs/languages) + # for a list of the currently supported language codes. + # Corresponds to the JSON property `languageCode` + # @return [String] + attr_accessor :language_code + + # *Optional* Maximum number of recognition hypotheses to be returned. + # Specifically, the maximum number of `SpeechRecognitionAlternative` messages + # within each `SpeechTranscription`. The server may return fewer than + # `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will + # return a maximum of one. If omitted, will return a maximum of one. + # Corresponds to the JSON property `maxAlternatives` + # @return [Fixnum] + attr_accessor :max_alternatives + + # *Optional* A means to provide context to assist the speech recognition. + # Corresponds to the JSON property `speechContexts` + # @return [Array] + attr_accessor :speech_contexts + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @audio_tracks = args[:audio_tracks] if args.key?(:audio_tracks) + @diarization_speaker_count = args[:diarization_speaker_count] if args.key?(:diarization_speaker_count) + @enable_automatic_punctuation = args[:enable_automatic_punctuation] if args.key?(:enable_automatic_punctuation) + @enable_speaker_diarization = args[:enable_speaker_diarization] if args.key?(:enable_speaker_diarization) + @enable_word_confidence = args[:enable_word_confidence] if args.key?(:enable_word_confidence) + @filter_profanity = args[:filter_profanity] if args.key?(:filter_profanity) + @language_code = args[:language_code] if args.key?(:language_code) + @max_alternatives = args[:max_alternatives] if args.key?(:max_alternatives) + @speech_contexts = args[:speech_contexts] if args.key?(:speech_contexts) + end + end + + # Annotations related to one detected OCR text snippet. This will contain the + # corresponding text, confidence value, and frame level information for each + # detection. + class GoogleCloudVideointelligenceV1p2beta1TextAnnotation + include Google::Apis::Core::Hashable + + # All video segments where OCR detected text appears. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + # The detected text. + # Corresponds to the JSON property `text` + # @return [String] + attr_accessor :text + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @segments = args[:segments] if args.key?(:segments) + @text = args[:text] if args.key?(:text) + end + end + + # Config for TEXT_DETECTION. + class GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig + include Google::Apis::Core::Hashable + + # Language hint can be specified if the language to be detected is known a + # priori. It can increase the accuracy of the detection. Language hint must + # be language code in BCP-47 format. + # Automatic language detection is performed if no hint is provided. + # Corresponds to the JSON property `languageHints` + # @return [Array] + attr_accessor :language_hints + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @language_hints = args[:language_hints] if args.key?(:language_hints) + end + end + + # Video frame level annotation results for text annotation (OCR). + # Contains information regarding timestamp and bounding box locations for the + # frames containing detected OCR text snippets. + class GoogleCloudVideointelligenceV1p2beta1TextFrame + include Google::Apis::Core::Hashable + + # Normalized bounding polygon for text (that might not be aligned with axis). + # Contains list of the corner points in clockwise order starting from + # top-left corner. For example, for a rectangular bounding box: + # When the text is horizontal it might look like: + # 0----1 + # | | + # 3----2 + # When it's clockwise rotated 180 degrees around the top-left corner it + # becomes: + # 2----3 + # | | + # 1----0 + # and the vertex order will still be (0, 1, 2, 3). Note that values can be less + # than 0, or greater than 1 due to trignometric calculations for location of + # the box. + # Corresponds to the JSON property `rotatedBoundingBox` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly] + attr_accessor :rotated_bounding_box + + # Timestamp of this frame. + # Corresponds to the JSON property `timeOffset` + # @return [String] + attr_accessor :time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @rotated_bounding_box = args[:rotated_bounding_box] if args.key?(:rotated_bounding_box) + @time_offset = args[:time_offset] if args.key?(:time_offset) + end + end + + # Video segment level annotation results for text detection. + class GoogleCloudVideointelligenceV1p2beta1TextSegment + include Google::Apis::Core::Hashable + + # Confidence for the track of detected text. It is calculated as the highest + # over all frames where OCR detected text appears. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Information related to the frames where OCR detected text appears. + # Corresponds to the JSON property `frames` + # @return [Array] + attr_accessor :frames + + # Video segment. + # Corresponds to the JSON property `segment` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment] + attr_accessor :segment + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @frames = args[:frames] if args.key?(:frames) + @segment = args[:segment] if args.key?(:segment) + end + end + + # Annotation progress for a single video. + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress + include Google::Apis::Core::Hashable + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Approximate percentage processed thus far. Guaranteed to be + # 100 when fully processed. + # Corresponds to the JSON property `progressPercent` + # @return [Fixnum] + attr_accessor :progress_percent + + # Time when the request was received. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # Time of the most recent update. + # Corresponds to the JSON property `updateTime` + # @return [String] + attr_accessor :update_time + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @progress_percent = args[:progress_percent] if args.key?(:progress_percent) + @start_time = args[:start_time] if args.key?(:start_time) + @update_time = args[:update_time] if args.key?(:update_time) + end + end + + # Annotation results for a single video. + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults + include Google::Apis::Core::Hashable + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + # Corresponds to the JSON property `error` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus] + attr_accessor :error + + # Explicit content annotation (based on per-frame visual signals only). + # If no explicit content has been detected in a frame, no annotations are + # present for that frame. + # Corresponds to the JSON property `explicitAnnotation` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation] + attr_accessor :explicit_annotation + + # Label annotations on frame level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `frameLabelAnnotations` + # @return [Array] + attr_accessor :frame_label_annotations + + # Video file location in + # [Google Cloud Storage](https://cloud.google.com/storage/). + # Corresponds to the JSON property `inputUri` + # @return [String] + attr_accessor :input_uri + + # Annotations for list of objects detected and tracked in video. + # Corresponds to the JSON property `objectAnnotations` + # @return [Array] + attr_accessor :object_annotations + + # Label annotations on video level or user specified segment level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `segmentLabelAnnotations` + # @return [Array] + attr_accessor :segment_label_annotations + + # Shot annotations. Each shot is represented as a video segment. + # Corresponds to the JSON property `shotAnnotations` + # @return [Array] + attr_accessor :shot_annotations + + # Label annotations on shot level. + # There is exactly one element for each unique label. + # Corresponds to the JSON property `shotLabelAnnotations` + # @return [Array] + attr_accessor :shot_label_annotations + + # Speech transcription. + # Corresponds to the JSON property `speechTranscriptions` + # @return [Array] + attr_accessor :speech_transcriptions + + # OCR text detection and tracking. + # Annotations for list of detected text snippets. Each will have list of + # frame information associated with it. + # Corresponds to the JSON property `textAnnotations` + # @return [Array] + attr_accessor :text_annotations + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @error = args[:error] if args.key?(:error) + @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation) + @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations) + @input_uri = args[:input_uri] if args.key?(:input_uri) + @object_annotations = args[:object_annotations] if args.key?(:object_annotations) + @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations) + @shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations) + @shot_label_annotations = args[:shot_label_annotations] if args.key?(:shot_label_annotations) + @speech_transcriptions = args[:speech_transcriptions] if args.key?(:speech_transcriptions) + @text_annotations = args[:text_annotations] if args.key?(:text_annotations) + end + end + + # Video context and/or feature-specific parameters. + class GoogleCloudVideointelligenceV1p2beta1VideoContext + include Google::Apis::Core::Hashable + + # Config for EXPLICIT_CONTENT_DETECTION. + # Corresponds to the JSON property `explicitContentDetectionConfig` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig] + attr_accessor :explicit_content_detection_config + + # Config for LABEL_DETECTION. + # Corresponds to the JSON property `labelDetectionConfig` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig] + attr_accessor :label_detection_config + + # Video segments to annotate. The segments may overlap and are not required + # to be contiguous or span the whole video. If unspecified, each video is + # treated as a single segment. + # Corresponds to the JSON property `segments` + # @return [Array] + attr_accessor :segments + + # Config for SHOT_CHANGE_DETECTION. + # Corresponds to the JSON property `shotChangeDetectionConfig` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig] + attr_accessor :shot_change_detection_config + + # Config for SPEECH_TRANSCRIPTION. + # Corresponds to the JSON property `speechTranscriptionConfig` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig] + attr_accessor :speech_transcription_config + + # Config for TEXT_DETECTION. + # Corresponds to the JSON property `textDetectionConfig` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig] + attr_accessor :text_detection_config + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @explicit_content_detection_config = args[:explicit_content_detection_config] if args.key?(:explicit_content_detection_config) + @label_detection_config = args[:label_detection_config] if args.key?(:label_detection_config) + @segments = args[:segments] if args.key?(:segments) + @shot_change_detection_config = args[:shot_change_detection_config] if args.key?(:shot_change_detection_config) + @speech_transcription_config = args[:speech_transcription_config] if args.key?(:speech_transcription_config) + @text_detection_config = args[:text_detection_config] if args.key?(:text_detection_config) + end + end + + # Video segment. + class GoogleCloudVideointelligenceV1p2beta1VideoSegment + include Google::Apis::Core::Hashable + + # Time-offset, relative to the beginning of the video, + # corresponding to the end of the segment (inclusive). + # Corresponds to the JSON property `endTimeOffset` + # @return [String] + attr_accessor :end_time_offset + + # Time-offset, relative to the beginning of the video, + # corresponding to the start of the segment (inclusive). + # Corresponds to the JSON property `startTimeOffset` + # @return [String] + attr_accessor :start_time_offset + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @end_time_offset = args[:end_time_offset] if args.key?(:end_time_offset) + @start_time_offset = args[:start_time_offset] if args.key?(:start_time_offset) + end + end + + # Word-specific information for recognized words. Word information is only + # included in the response when certain request parameters are set, such + # as `enable_word_time_offsets`. + class GoogleCloudVideointelligenceV1p2beta1WordInfo + include Google::Apis::Core::Hashable + + # Output only. The confidence estimate between 0.0 and 1.0. A higher number + # indicates an estimated greater likelihood that the recognized words are + # correct. This field is set only for the top alternative. + # This field is not guaranteed to be accurate and users should not rely on it + # to be always provided. + # The default of 0.0 is a sentinel value indicating `confidence` was not set. + # Corresponds to the JSON property `confidence` + # @return [Float] + attr_accessor :confidence + + # Time offset relative to the beginning of the audio, and + # corresponding to the end of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `endTime` + # @return [String] + attr_accessor :end_time + + # Output only. A distinct integer value is assigned for every speaker within + # the audio. This field specifies which one of those speakers was detected to + # have spoken this word. Value ranges from 1 up to diarization_speaker_count, + # and is only set if speaker diarization is enabled. + # Corresponds to the JSON property `speakerTag` + # @return [Fixnum] + attr_accessor :speaker_tag + + # Time offset relative to the beginning of the audio, and + # corresponding to the start of the spoken word. This field is only set if + # `enable_word_time_offsets=true` and only in the top hypothesis. This is an + # experimental feature and the accuracy of the time offset can vary. + # Corresponds to the JSON property `startTime` + # @return [String] + attr_accessor :start_time + + # The word corresponding to this set of information. + # Corresponds to the JSON property `word` + # @return [String] + attr_accessor :word + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @confidence = args[:confidence] if args.key?(:confidence) + @end_time = args[:end_time] if args.key?(:end_time) + @speaker_tag = args[:speaker_tag] if args.key?(:speaker_tag) + @start_time = args[:start_time] if args.key?(:start_time) + @word = args[:word] if args.key?(:word) + end + end + + # This resource represents a long-running operation that is the result of a + # network API call. + class GoogleLongrunningOperation + include Google::Apis::Core::Hashable + + # If the value is `false`, it means the operation is still in progress. + # If `true`, the operation is completed, and either `error` or `response` is + # available. + # Corresponds to the JSON property `done` + # @return [Boolean] + attr_accessor :done + alias_method :done?, :done + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + # Corresponds to the JSON property `error` + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus] + attr_accessor :error + + # Service-specific metadata associated with the operation. It typically + # contains progress information and common metadata such as create time. + # Some services might not provide such metadata. Any method that returns a + # long-running operation should document the metadata type, if any. + # Corresponds to the JSON property `metadata` + # @return [Hash] + attr_accessor :metadata + + # The server-assigned name, which is only unique within the same service that + # originally returns it. If you use the default HTTP mapping, the + # `name` should have the format of `operations/some/unique/name`. + # Corresponds to the JSON property `name` + # @return [String] + attr_accessor :name + + # The normal response of the operation in case of success. If the original + # method returns no data on success, such as `Delete`, the response is + # `google.protobuf.Empty`. If the original method is standard + # `Get`/`Create`/`Update`, the response should be the resource. For other + # methods, the response should have the type `XxxResponse`, where `Xxx` + # is the original method name. For example, if the original method name + # is `TakeSnapshot()`, the inferred response type is + # `TakeSnapshotResponse`. + # Corresponds to the JSON property `response` + # @return [Hash] + attr_accessor :response + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @done = args[:done] if args.key?(:done) + @error = args[:error] if args.key?(:error) + @metadata = args[:metadata] if args.key?(:metadata) + @name = args[:name] if args.key?(:name) + @response = args[:response] if args.key?(:response) + end + end + + # The `Status` type defines a logical error model that is suitable for different + # programming environments, including REST APIs and RPC APIs. It is used by + # [gRPC](https://github.com/grpc). The error model is designed to be: + # - Simple to use and understand for most users + # - Flexible enough to meet unexpected needs + # # Overview + # The `Status` message contains three pieces of data: error code, error message, + # and error details. The error code should be an enum value of + # google.rpc.Code, but it may accept additional error codes if needed. The + # error message should be a developer-facing English message that helps + # developers *understand* and *resolve* the error. If a localized user-facing + # error message is needed, put the localized message in the error details or + # localize it in the client. The optional error details may contain arbitrary + # information about the error. There is a predefined set of error detail types + # in the package `google.rpc` that can be used for common error conditions. + # # Language mapping + # The `Status` message is the logical representation of the error model, but it + # is not necessarily the actual wire format. When the `Status` message is + # exposed in different client libraries and different wire protocols, it can be + # mapped differently. For example, it will likely be mapped to some exceptions + # in Java, but more likely mapped to some error codes in C. + # # Other uses + # The error model and the `Status` message can be used in a variety of + # environments, either with or without APIs, to provide a + # consistent developer experience across different environments. + # Example uses of this error model include: + # - Partial errors. If a service needs to return partial errors to the client, + # it may embed the `Status` in the normal response to indicate the partial + # errors. + # - Workflow errors. A typical workflow has multiple steps. Each step may + # have a `Status` message for error reporting. + # - Batch operations. If a client uses batch request and batch response, the + # `Status` message should be used directly inside batch response, one for + # each error sub-response. + # - Asynchronous operations. If an API call embeds asynchronous operation + # results in its response, the status of those operations should be + # represented directly using the `Status` message. + # - Logging. If some API errors are stored in logs, the message `Status` could + # be used directly after any stripping needed for security/privacy reasons. + class GoogleRpcStatus + include Google::Apis::Core::Hashable + + # The status code, which should be an enum value of google.rpc.Code. + # Corresponds to the JSON property `code` + # @return [Fixnum] + attr_accessor :code + + # A list of messages that carry the error details. There is a common set of + # message types for APIs to use. + # Corresponds to the JSON property `details` + # @return [Array>] + attr_accessor :details + + # A developer-facing error message, which should be in English. Any + # user-facing error message should be localized and sent in the + # google.rpc.Status.details field, or localized by the client. + # Corresponds to the JSON property `message` + # @return [String] + attr_accessor :message + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @code = args[:code] if args.key?(:code) + @details = args[:details] if args.key?(:details) + @message = args[:message] if args.key?(:message) + end + end + end + end +end diff --git a/generated/google/apis/videointelligence_v1p2beta1/representations.rb b/generated/google/apis/videointelligence_v1p2beta1/representations.rb new file mode 100644 index 000000000..047480a9c --- /dev/null +++ b/generated/google/apis/videointelligence_v1p2beta1/representations.rb @@ -0,0 +1,1221 @@ +# Copyright 2015 Google Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +require 'date' +require 'google/apis/core/base_service' +require 'google/apis/core/json_representation' +require 'google/apis/core/hashable' +require 'google/apis/errors' + +module Google + module Apis + module VideointelligenceV1p2beta1 + + class GoogleCloudVideointelligenceV1AnnotateVideoProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1AnnotateVideoResponse + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1Entity + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1ExplicitContentAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1ExplicitContentFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1LabelAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1LabelFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1LabelSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1SpeechTranscription + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1VideoAnnotationProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1VideoAnnotationResults + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1VideoSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1WordInfo + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2Entity + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2ExplicitContentFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2LabelAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2LabelFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2LabelSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2SpeechTranscription + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2VideoAnnotationResults + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2VideoSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1beta2WordInfo + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoResponse + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1Entity + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1LabelAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1LabelFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1LabelSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1SpeechTranscription + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1VideoSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p1beta1WordInfo + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoRequest + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1Entity + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1LabelFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1LabelSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechContext + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1TextAnnotation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1TextFrame + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1TextSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1VideoContext + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1VideoSegment + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1p2beta1WordInfo + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleLongrunningOperation + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleRpcStatus + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + + class GoogleCloudVideointelligenceV1AnnotateVideoProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_progress, as: 'annotationProgress', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoAnnotationProgress, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoAnnotationProgress::Representation + + end + end + + class GoogleCloudVideointelligenceV1AnnotateVideoResponse + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_results, as: 'annotationResults', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoAnnotationResults, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoAnnotationResults::Representation + + end + end + + class GoogleCloudVideointelligenceV1Entity + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :description, as: 'description' + property :entity_id, as: 'entityId' + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1ExplicitContentAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1ExplicitContentFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1ExplicitContentFrame::Representation + + end + end + + class GoogleCloudVideointelligenceV1ExplicitContentFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :pornography_likelihood, as: 'pornographyLikelihood' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1LabelAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :category_entities, as: 'categoryEntities', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1Entity::Representation + + property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1Entity::Representation + + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelFrame::Representation + + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1LabelFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1LabelSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :transcript, as: 'transcript' + collection :words, as: 'words', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1WordInfo, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1WordInfo::Representation + + end + end + + class GoogleCloudVideointelligenceV1SpeechTranscription + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative::Representation + + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1VideoAnnotationProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :input_uri, as: 'inputUri' + property :progress_percent, as: 'progressPercent' + property :start_time, as: 'startTime' + property :update_time, as: 'updateTime' + end + end + + class GoogleCloudVideointelligenceV1VideoAnnotationResults + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :error, as: 'error', class: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus::Representation + + property :explicit_annotation, as: 'explicitAnnotation', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1ExplicitContentAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1ExplicitContentAnnotation::Representation + + collection :frame_label_annotations, as: 'frameLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation::Representation + + property :input_uri, as: 'inputUri' + collection :segment_label_annotations, as: 'segmentLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation::Representation + + collection :shot_annotations, as: 'shotAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1VideoSegment::Representation + + collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1LabelAnnotation::Representation + + collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1SpeechTranscription, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1SpeechTranscription::Representation + + end + end + + class GoogleCloudVideointelligenceV1VideoSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :end_time_offset, as: 'endTimeOffset' + property :start_time_offset, as: 'startTimeOffset' + end + end + + class GoogleCloudVideointelligenceV1WordInfo + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :end_time, as: 'endTime' + property :speaker_tag, as: 'speakerTag' + property :start_time, as: 'startTime' + property :word, as: 'word' + end + end + + class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_progress, as: 'annotationProgress', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_results, as: 'annotationResults', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoAnnotationResults, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoAnnotationResults::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2Entity + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :description, as: 'description' + property :entity_id, as: 'entityId' + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentFrame::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2ExplicitContentFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :pornography_likelihood, as: 'pornographyLikelihood' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1beta2LabelAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :category_entities, as: 'categoryEntities', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2Entity::Representation + + property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2Entity::Representation + + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelFrame::Representation + + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2LabelFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1beta2LabelSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :transcript, as: 'transcript' + collection :words, as: 'words', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2WordInfo, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2WordInfo::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2SpeechTranscription + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative::Representation + + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :input_uri, as: 'inputUri' + property :progress_percent, as: 'progressPercent' + property :start_time, as: 'startTime' + property :update_time, as: 'updateTime' + end + end + + class GoogleCloudVideointelligenceV1beta2VideoAnnotationResults + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :error, as: 'error', class: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus::Representation + + property :explicit_annotation, as: 'explicitAnnotation', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation::Representation + + collection :frame_label_annotations, as: 'frameLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation::Representation + + property :input_uri, as: 'inputUri' + collection :segment_label_annotations, as: 'segmentLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation::Representation + + collection :shot_annotations, as: 'shotAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2VideoSegment::Representation + + collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation::Representation + + collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2SpeechTranscription, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1beta2SpeechTranscription::Representation + + end + end + + class GoogleCloudVideointelligenceV1beta2VideoSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :end_time_offset, as: 'endTimeOffset' + property :start_time_offset, as: 'startTimeOffset' + end + end + + class GoogleCloudVideointelligenceV1beta2WordInfo + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :end_time, as: 'endTime' + property :speaker_tag, as: 'speakerTag' + property :start_time, as: 'startTime' + property :word, as: 'word' + end + end + + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_progress, as: 'annotationProgress', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoResponse + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_results, as: 'annotationResults', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1Entity + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :description, as: 'description' + property :entity_id, as: 'entityId' + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :pornography_likelihood, as: 'pornographyLikelihood' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p1beta1LabelAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :category_entities, as: 'categoryEntities', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1Entity::Representation + + property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1Entity::Representation + + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelFrame::Representation + + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1LabelFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p1beta1LabelSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :transcript, as: 'transcript' + collection :words, as: 'words', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1WordInfo, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1WordInfo::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1SpeechTranscription + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative::Representation + + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :input_uri, as: 'inputUri' + property :progress_percent, as: 'progressPercent' + property :start_time, as: 'startTime' + property :update_time, as: 'updateTime' + end + end + + class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :error, as: 'error', class: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus::Representation + + property :explicit_annotation, as: 'explicitAnnotation', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation::Representation + + collection :frame_label_annotations, as: 'frameLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation::Representation + + property :input_uri, as: 'inputUri' + collection :segment_label_annotations, as: 'segmentLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation::Representation + + collection :shot_annotations, as: 'shotAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment::Representation + + collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation::Representation + + collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1SpeechTranscription, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p1beta1SpeechTranscription::Representation + + end + end + + class GoogleCloudVideointelligenceV1p1beta1VideoSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :end_time_offset, as: 'endTimeOffset' + property :start_time_offset, as: 'startTimeOffset' + end + end + + class GoogleCloudVideointelligenceV1p1beta1WordInfo + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :end_time, as: 'endTime' + property :speaker_tag, as: 'speakerTag' + property :start_time, as: 'startTime' + property :word, as: 'word' + end + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_progress, as: 'annotationProgress', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoRequest + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :features, as: 'features' + property :input_content, :base64 => true, as: 'inputContent' + property :input_uri, as: 'inputUri' + property :location_id, as: 'locationId' + property :output_uri, as: 'outputUri' + property :video_context, as: 'videoContext', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoContext, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoContext::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :annotation_results, as: 'annotationResults', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1Entity + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :description, as: 'description' + property :entity_id, as: 'entityId' + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :model, as: 'model' + end + end + + class GoogleCloudVideointelligenceV1p2beta1ExplicitContentFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :pornography_likelihood, as: 'pornographyLikelihood' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :category_entities, as: 'categoryEntities', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation + + property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation + + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelFrame::Representation + + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :label_detection_mode, as: 'labelDetectionMode' + property :model, as: 'model' + property :stationary_camera, as: 'stationaryCamera' + end + end + + class GoogleCloudVideointelligenceV1p2beta1LabelFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p2beta1LabelSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :bottom, as: 'bottom' + property :left, as: 'left' + property :right, as: 'right' + property :top, as: 'top' + end + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :vertices, as: 'vertices', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :x, as: 'x' + property :y, as: 'y' + end + end + + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :entity, as: 'entity', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1Entity::Representation + + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame::Representation + + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation + + property :track_id, :numeric_string => true, as: 'trackId' + end + end + + class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :normalized_bounding_box, as: 'normalizedBoundingBox', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox::Representation + + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :model, as: 'model' + end + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechContext + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :phrases, as: 'phrases' + end + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :transcript, as: 'transcript' + collection :words, as: 'words', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1WordInfo, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1WordInfo::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :alternatives, as: 'alternatives', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative::Representation + + property :language_code, as: 'languageCode' + end + end + + class GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :audio_tracks, as: 'audioTracks' + property :diarization_speaker_count, as: 'diarizationSpeakerCount' + property :enable_automatic_punctuation, as: 'enableAutomaticPunctuation' + property :enable_speaker_diarization, as: 'enableSpeakerDiarization' + property :enable_word_confidence, as: 'enableWordConfidence' + property :filter_profanity, as: 'filterProfanity' + property :language_code, as: 'languageCode' + property :max_alternatives, as: 'maxAlternatives' + collection :speech_contexts, as: 'speechContexts', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechContext, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechContext::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1TextAnnotation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextSegment::Representation + + property :text, as: 'text' + end + end + + class GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig + # @private + class Representation < Google::Apis::Core::JsonRepresentation + collection :language_hints, as: 'languageHints' + end + end + + class GoogleCloudVideointelligenceV1p2beta1TextFrame + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :rotated_bounding_box, as: 'rotatedBoundingBox', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly::Representation + + property :time_offset, as: 'timeOffset' + end + end + + class GoogleCloudVideointelligenceV1p2beta1TextSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + collection :frames, as: 'frames', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextFrame, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextFrame::Representation + + property :segment, as: 'segment', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :input_uri, as: 'inputUri' + property :progress_percent, as: 'progressPercent' + property :start_time, as: 'startTime' + property :update_time, as: 'updateTime' + end + end + + class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :error, as: 'error', class: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus::Representation + + property :explicit_annotation, as: 'explicitAnnotation', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation::Representation + + collection :frame_label_annotations, as: 'frameLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation + + property :input_uri, as: 'inputUri' + collection :object_annotations, as: 'objectAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation::Representation + + collection :segment_label_annotations, as: 'segmentLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation + + collection :shot_annotations, as: 'shotAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation + + collection :shot_label_annotations, as: 'shotLabelAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation::Representation + + collection :speech_transcriptions, as: 'speechTranscriptions', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription::Representation + + collection :text_annotations, as: 'textAnnotations', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1VideoContext + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :explicit_content_detection_config, as: 'explicitContentDetectionConfig', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentDetectionConfig::Representation + + property :label_detection_config, as: 'labelDetectionConfig', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1LabelDetectionConfig::Representation + + collection :segments, as: 'segments', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment::Representation + + property :shot_change_detection_config, as: 'shotChangeDetectionConfig', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1ShotChangeDetectionConfig::Representation + + property :speech_transcription_config, as: 'speechTranscriptionConfig', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscriptionConfig::Representation + + property :text_detection_config, as: 'textDetectionConfig', class: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1TextDetectionConfig::Representation + + end + end + + class GoogleCloudVideointelligenceV1p2beta1VideoSegment + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :end_time_offset, as: 'endTimeOffset' + property :start_time_offset, as: 'startTimeOffset' + end + end + + class GoogleCloudVideointelligenceV1p2beta1WordInfo + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :confidence, as: 'confidence' + property :end_time, as: 'endTime' + property :speaker_tag, as: 'speakerTag' + property :start_time, as: 'startTime' + property :word, as: 'word' + end + end + + class GoogleLongrunningOperation + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :done, as: 'done' + property :error, as: 'error', class: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus, decorator: Google::Apis::VideointelligenceV1p2beta1::GoogleRpcStatus::Representation + + hash :metadata, as: 'metadata' + property :name, as: 'name' + hash :response, as: 'response' + end + end + + class GoogleRpcStatus + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :code, as: 'code' + collection :details, as: 'details' + property :message, as: 'message' + end + end + end + end +end diff --git a/generated/google/apis/videointelligence_v1p2beta1/service.rb b/generated/google/apis/videointelligence_v1p2beta1/service.rb new file mode 100644 index 000000000..b6b3ea8fb --- /dev/null +++ b/generated/google/apis/videointelligence_v1p2beta1/service.rb @@ -0,0 +1,93 @@ +# Copyright 2015 Google Inc. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +require 'google/apis/core/base_service' +require 'google/apis/core/json_representation' +require 'google/apis/core/hashable' +require 'google/apis/errors' + +module Google + module Apis + module VideointelligenceV1p2beta1 + # Cloud Video Intelligence API + # + # Detects objects, explicit content, and scene changes in videos. It also + # specifies the region for annotation and transcribes speech to text. + # + # @example + # require 'google/apis/videointelligence_v1p2beta1' + # + # Videointelligence = Google::Apis::VideointelligenceV1p2beta1 # Alias the module + # service = Videointelligence::CloudVideoIntelligenceService.new + # + # @see https://cloud.google.com/video-intelligence/docs/ + class CloudVideoIntelligenceService < Google::Apis::Core::BaseService + # @return [String] + # API key. Your API key identifies your project and provides you with API access, + # quota, and reports. Required unless you provide an OAuth 2.0 token. + attr_accessor :key + + # @return [String] + # Available to use for quota purposes for server-side applications. Can be any + # arbitrary string assigned to a user, but should not exceed 40 characters. + attr_accessor :quota_user + + def initialize + super('https://videointelligence.googleapis.com/', '') + @batch_path = 'batch' + end + + # Performs asynchronous video annotation. Progress and results can be + # retrieved through the `google.longrunning.Operations` interface. + # `Operation.metadata` contains `AnnotateVideoProgress` (progress). + # `Operation.response` contains `AnnotateVideoResponse` (results). + # @param [Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1AnnotateVideoRequest] google_cloud_videointelligence_v1p2beta1_annotate_video_request_object + # @param [String] fields + # Selector specifying which fields to include in a partial response. + # @param [String] quota_user + # Available to use for quota purposes for server-side applications. Can be any + # arbitrary string assigned to a user, but should not exceed 40 characters. + # @param [Google::Apis::RequestOptions] options + # Request-specific options + # + # @yield [result, err] Result & error if block supplied + # @yieldparam result [Google::Apis::VideointelligenceV1p2beta1::GoogleLongrunningOperation] parsed result object + # @yieldparam err [StandardError] error object if request failed + # + # @return [Google::Apis::VideointelligenceV1p2beta1::GoogleLongrunningOperation] + # + # @raise [Google::Apis::ServerError] An error occurred on the server and the request can be retried + # @raise [Google::Apis::ClientError] The request is invalid and should not be retried without modification + # @raise [Google::Apis::AuthorizationError] Authorization is required + def annotate_video(google_cloud_videointelligence_v1p2beta1_annotate_video_request_object = nil, fields: nil, quota_user: nil, options: nil, &block) + command = make_simple_command(:post, 'v1p2beta1/videos:annotate', options) + command.request_representation = Google::Apis::VideointelligenceV1p2beta1::GoogleCloudVideointelligenceV1p2beta1AnnotateVideoRequest::Representation + command.request_object = google_cloud_videointelligence_v1p2beta1_annotate_video_request_object + command.response_representation = Google::Apis::VideointelligenceV1p2beta1::GoogleLongrunningOperation::Representation + command.response_class = Google::Apis::VideointelligenceV1p2beta1::GoogleLongrunningOperation + command.query['fields'] = fields unless fields.nil? + command.query['quotaUser'] = quota_user unless quota_user.nil? + execute_or_queue_command(command, &block) + end + + protected + + def apply_command_defaults(command) + command.query['key'] = key unless key.nil? + command.query['quotaUser'] = quota_user unless quota_user.nil? + end + end + end + end +end diff --git a/generated/google/apis/youtube_v3.rb b/generated/google/apis/youtube_v3.rb index 7eef6979f..8cfa991cf 100644 --- a/generated/google/apis/youtube_v3.rb +++ b/generated/google/apis/youtube_v3.rb @@ -26,7 +26,7 @@ module Google # @see https://developers.google.com/youtube/v3 module YoutubeV3 VERSION = 'V3' - REVISION = '20180511' + REVISION = '20190128' # Manage your YouTube account AUTH_YOUTUBE = 'https://www.googleapis.com/auth/youtube' diff --git a/generated/google/apis/youtube_v3/classes.rb b/generated/google/apis/youtube_v3/classes.rb index b8e418cbd..227d75c16 100644 --- a/generated/google/apis/youtube_v3/classes.rb +++ b/generated/google/apis/youtube_v3/classes.rb @@ -1003,16 +1003,6 @@ module Google attr_accessor :copyright_strikes_good_standing alias_method :copyright_strikes_good_standing?, :copyright_strikes_good_standing - # Describes the general state of the channel. This field will always show if - # there are any issues whatsoever with the channel. Currently this field - # represents the result of the logical and operation over the community - # guidelines good standing, the copyright strikes good standing and the content - # ID claims good standing, but this may change in the future. - # Corresponds to the JSON property `overallGoodStanding` - # @return [Boolean] - attr_accessor :overall_good_standing - alias_method :overall_good_standing?, :overall_good_standing - def initialize(**args) update!(**args) end @@ -1022,7 +1012,6 @@ module Google @community_guidelines_good_standing = args[:community_guidelines_good_standing] if args.key?(:community_guidelines_good_standing) @content_id_claims_good_standing = args[:content_id_claims_good_standing] if args.key?(:content_id_claims_good_standing) @copyright_strikes_good_standing = args[:copyright_strikes_good_standing] if args.key?(:copyright_strikes_good_standing) - @overall_good_standing = args[:overall_good_standing] if args.key?(:overall_good_standing) end end @@ -4367,6 +4356,12 @@ module Google # @return [Google::Apis::YoutubeV3::LiveChatSuperChatDetails] attr_accessor :super_chat_details + # Details about the Super Sticker event, this is only set if the type is ' + # superStickerEvent'. + # Corresponds to the JSON property `superStickerDetails` + # @return [Google::Apis::YoutubeV3::LiveChatSuperStickerDetails] + attr_accessor :super_sticker_details + # Details about the text message, this is only set if the type is ' # textMessageEvent'. # Corresponds to the JSON property `textMessageDetails` @@ -4403,6 +4398,7 @@ module Google @poll_voted_details = args[:poll_voted_details] if args.key?(:poll_voted_details) @published_at = args[:published_at] if args.key?(:published_at) @super_chat_details = args[:super_chat_details] if args.key?(:super_chat_details) + @super_sticker_details = args[:super_sticker_details] if args.key?(:super_sticker_details) @text_message_details = args[:text_message_details] if args.key?(:text_message_details) @type = args[:type] if args.key?(:type) @user_banned_details = args[:user_banned_details] if args.key?(:user_banned_details) @@ -4695,8 +4691,8 @@ module Google # @return [String] attr_accessor :currency - # The tier in which the amount belongs to. Lower amounts belong to lower tiers. - # Starts at 1. + # The tier in which the amount belongs. Lower amounts belong to lower tiers. The + # lowest tier is 1. # Corresponds to the JSON property `tier` # @return [Fixnum] attr_accessor :tier @@ -4720,6 +4716,50 @@ module Google end end + # + class LiveChatSuperStickerDetails + include Google::Apis::Core::Hashable + + # A rendered string that displays the fund amount and currency to the user. + # Corresponds to the JSON property `amountDisplayString` + # @return [String] + attr_accessor :amount_display_string + + # The amount purchased by the user, in micros (1,750,000 micros = 1.75). + # Corresponds to the JSON property `amountMicros` + # @return [Fixnum] + attr_accessor :amount_micros + + # The currency in which the purchase was made. + # Corresponds to the JSON property `currency` + # @return [String] + attr_accessor :currency + + # Information about the Super Sticker. + # Corresponds to the JSON property `superStickerMetadata` + # @return [Google::Apis::YoutubeV3::SuperStickerMetadata] + attr_accessor :super_sticker_metadata + + # The tier in which the amount belongs. Lower amounts belong to lower tiers. The + # lowest tier is 1. + # Corresponds to the JSON property `tier` + # @return [Fixnum] + attr_accessor :tier + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @amount_display_string = args[:amount_display_string] if args.key?(:amount_display_string) + @amount_micros = args[:amount_micros] if args.key?(:amount_micros) + @currency = args[:currency] if args.key?(:currency) + @super_sticker_metadata = args[:super_sticker_metadata] if args.key?(:super_sticker_metadata) + @tier = args[:tier] if args.key?(:tier) + end + end + # class LiveChatTextMessageDetails include Google::Apis::Core::Hashable @@ -6692,6 +6732,12 @@ module Google attr_accessor :is_super_chat_for_good alias_method :is_super_chat_for_good?, :is_super_chat_for_good + # True if this event is a Super Sticker event. + # Corresponds to the JSON property `isSuperStickerEvent` + # @return [Boolean] + attr_accessor :is_super_sticker_event + alias_method :is_super_sticker_event?, :is_super_sticker_event + # The tier for the paid message, which is based on the amount of money spent to # purchase the message. # Corresponds to the JSON property `messageType` @@ -6703,6 +6749,12 @@ module Google # @return [Google::Apis::YoutubeV3::Nonprofit] attr_accessor :nonprofit + # If this event is a Super Sticker event, this field will contain metadata about + # the Super Sticker. + # Corresponds to the JSON property `superStickerMetadata` + # @return [Google::Apis::YoutubeV3::SuperStickerMetadata] + attr_accessor :super_sticker_metadata + # Details about the supporter. # Corresponds to the JSON property `supporterDetails` # @return [Google::Apis::YoutubeV3::ChannelProfileDetails] @@ -6721,12 +6773,47 @@ module Google @currency = args[:currency] if args.key?(:currency) @display_string = args[:display_string] if args.key?(:display_string) @is_super_chat_for_good = args[:is_super_chat_for_good] if args.key?(:is_super_chat_for_good) + @is_super_sticker_event = args[:is_super_sticker_event] if args.key?(:is_super_sticker_event) @message_type = args[:message_type] if args.key?(:message_type) @nonprofit = args[:nonprofit] if args.key?(:nonprofit) + @super_sticker_metadata = args[:super_sticker_metadata] if args.key?(:super_sticker_metadata) @supporter_details = args[:supporter_details] if args.key?(:supporter_details) end end + # + class SuperStickerMetadata + include Google::Apis::Core::Hashable + + # Internationalized alt text that describes the sticker image and any animation + # associated with it. + # Corresponds to the JSON property `altText` + # @return [String] + attr_accessor :alt_text + + # Specifies the localization language in which the alt text is returned. + # Corresponds to the JSON property `altTextLanguage` + # @return [String] + attr_accessor :alt_text_language + + # Unique identifier of the Super Sticker. This is a shorter form of the alt_text + # that includes pack name and a recognizable characteristic of the sticker. + # Corresponds to the JSON property `stickerId` + # @return [String] + attr_accessor :sticker_id + + def initialize(**args) + update!(**args) + end + + # Update properties of this object + def update!(**args) + @alt_text = args[:alt_text] if args.key?(:alt_text) + @alt_text_language = args[:alt_text_language] if args.key?(:alt_text_language) + @sticker_id = args[:sticker_id] if args.key?(:sticker_id) + end + end + # A thumbnail is an image representing a YouTube resource. class Thumbnail include Google::Apis::Core::Hashable diff --git a/generated/google/apis/youtube_v3/representations.rb b/generated/google/apis/youtube_v3/representations.rb index d52b8355d..b1bb27187 100644 --- a/generated/google/apis/youtube_v3/representations.rb +++ b/generated/google/apis/youtube_v3/representations.rb @@ -574,6 +574,12 @@ module Google include Google::Apis::Core::JsonObjectSupport end + class LiveChatSuperStickerDetails + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + class LiveChatTextMessageDetails class Representation < Google::Apis::Core::JsonRepresentation; end @@ -844,6 +850,12 @@ module Google include Google::Apis::Core::JsonObjectSupport end + class SuperStickerMetadata + class Representation < Google::Apis::Core::JsonRepresentation; end + + include Google::Apis::Core::JsonObjectSupport + end + class Thumbnail class Representation < Google::Apis::Core::JsonRepresentation; end @@ -1345,7 +1357,6 @@ module Google property :community_guidelines_good_standing, as: 'communityGuidelinesGoodStanding' property :content_id_claims_good_standing, as: 'contentIdClaimsGoodStanding' property :copyright_strikes_good_standing, as: 'copyrightStrikesGoodStanding' - property :overall_good_standing, as: 'overallGoodStanding' end end @@ -2180,6 +2191,8 @@ module Google property :super_chat_details, as: 'superChatDetails', class: Google::Apis::YoutubeV3::LiveChatSuperChatDetails, decorator: Google::Apis::YoutubeV3::LiveChatSuperChatDetails::Representation + property :super_sticker_details, as: 'superStickerDetails', class: Google::Apis::YoutubeV3::LiveChatSuperStickerDetails, decorator: Google::Apis::YoutubeV3::LiveChatSuperStickerDetails::Representation + property :text_message_details, as: 'textMessageDetails', class: Google::Apis::YoutubeV3::LiveChatTextMessageDetails, decorator: Google::Apis::YoutubeV3::LiveChatTextMessageDetails::Representation property :type, as: 'type' @@ -2280,6 +2293,18 @@ module Google end end + class LiveChatSuperStickerDetails + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :amount_display_string, as: 'amountDisplayString' + property :amount_micros, :numeric_string => true, as: 'amountMicros' + property :currency, as: 'currency' + property :super_sticker_metadata, as: 'superStickerMetadata', class: Google::Apis::YoutubeV3::SuperStickerMetadata, decorator: Google::Apis::YoutubeV3::SuperStickerMetadata::Representation + + property :tier, as: 'tier' + end + end + class LiveChatTextMessageDetails # @private class Representation < Google::Apis::Core::JsonRepresentation @@ -2820,14 +2845,26 @@ module Google property :currency, as: 'currency' property :display_string, as: 'displayString' property :is_super_chat_for_good, as: 'isSuperChatForGood' + property :is_super_sticker_event, as: 'isSuperStickerEvent' property :message_type, as: 'messageType' property :nonprofit, as: 'nonprofit', class: Google::Apis::YoutubeV3::Nonprofit, decorator: Google::Apis::YoutubeV3::Nonprofit::Representation + property :super_sticker_metadata, as: 'superStickerMetadata', class: Google::Apis::YoutubeV3::SuperStickerMetadata, decorator: Google::Apis::YoutubeV3::SuperStickerMetadata::Representation + property :supporter_details, as: 'supporterDetails', class: Google::Apis::YoutubeV3::ChannelProfileDetails, decorator: Google::Apis::YoutubeV3::ChannelProfileDetails::Representation end end + class SuperStickerMetadata + # @private + class Representation < Google::Apis::Core::JsonRepresentation + property :alt_text, as: 'altText' + property :alt_text_language, as: 'altTextLanguage' + property :sticker_id, as: 'stickerId' + end + end + class Thumbnail # @private class Representation < Google::Apis::Core::JsonRepresentation