Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(272)

Side by Side Diff: generated/googleapis_beta/lib/videointelligence/v1beta1.dart

Issue 3006323002: Api-Roll 54: 2017-09-11 (Closed)
Patch Set: use 2.0.0-dev.infinity sdk constraint in pubspecs Created 3 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 // This is a generated file (see the discoveryapis_generator project). 1 // This is a generated file (see the discoveryapis_generator project).
2 2
3 library googleapis_beta.videointelligence.v1beta1; 3 library googleapis_beta.videointelligence.v1beta1;
4 4
5 import 'dart:core' as core; 5 import 'dart:core' as core;
6 import 'dart:async' as async; 6 import 'dart:async' as async;
7 import 'dart:convert' as convert; 7 import 'dart:convert' as convert;
8 8
9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons; 9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons;
10 import 'package:http/http.dart' as http; 10 import 'package:http/http.dart' as http;
11 11
12 export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show 12 export 'package:_discoveryapis_commons/_discoveryapis_commons.dart'
13 ApiRequestError, DetailedApiRequestError; 13 show ApiRequestError, DetailedApiRequestError;
14 14
15 const core.String USER_AGENT = 'dart-api-client videointelligence/v1beta1'; 15 const core.String USER_AGENT = 'dart-api-client videointelligence/v1beta1';
16 16
17 /** Google Cloud Video Intelligence API. */ 17 /// Google Cloud Video Intelligence API.
18 class VideointelligenceApi { 18 class VideointelligenceApi {
19 /** View and manage your data across Google Cloud Platform services */ 19 /// View and manage your data across Google Cloud Platform services
20 static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platf orm"; 20 static const CloudPlatformScope =
21 21 "https://www.googleapis.com/auth/cloud-platform";
22 22
23 final commons.ApiRequester _requester; 23 final commons.ApiRequester _requester;
24 24
25 VideosResourceApi get videos => new VideosResourceApi(_requester); 25 VideosResourceApi get videos => new VideosResourceApi(_requester);
26 26
27 VideointelligenceApi(http.Client client, {core.String rootUrl: "https://videoi ntelligence.googleapis.com/", core.String servicePath: ""}) : 27 VideointelligenceApi(http.Client client,
28 _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_A GENT); 28 {core.String rootUrl: "https://videointelligence.googleapis.com/",
29 core.String servicePath: ""})
30 : _requester =
31 new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
29 } 32 }
30 33
31
32 class VideosResourceApi { 34 class VideosResourceApi {
33 final commons.ApiRequester _requester; 35 final commons.ApiRequester _requester;
34 36
35 VideosResourceApi(commons.ApiRequester client) : 37 VideosResourceApi(commons.ApiRequester client) : _requester = client;
36 _requester = client;
37 38
38 /** 39 /// Performs asynchronous video annotation. Progress and results can be
39 * Performs asynchronous video annotation. Progress and results can be 40 /// retrieved through the `google.longrunning.Operations` interface.
40 * retrieved through the `google.longrunning.Operations` interface. 41 /// `Operation.metadata` contains `AnnotateVideoProgress` (progress).
41 * `Operation.metadata` contains `AnnotateVideoProgress` (progress). 42 /// `Operation.response` contains `AnnotateVideoResponse` (results).
42 * `Operation.response` contains `AnnotateVideoResponse` (results). 43 ///
43 * 44 /// [request] - The metadata request object.
44 * [request] - The metadata request object. 45 ///
45 * 46 /// Request parameters:
46 * Request parameters: 47 ///
47 * 48 /// Completes with a [GoogleLongrunningOperation].
48 * Completes with a [GoogleLongrunningOperation]. 49 ///
49 * 50 /// Completes with a [commons.ApiRequestError] if the API endpoint returned
50 * Completes with a [commons.ApiRequestError] if the API endpoint returned an 51 /// an error.
51 * error. 52 ///
52 * 53 /// If the used [http.Client] completes with an error when making a REST
53 * If the used [http.Client] completes with an error when making a REST call, 54 /// call, this method will complete with the same error.
54 * this method will complete with the same error. 55 async.Future<GoogleLongrunningOperation> annotate(
55 */ 56 GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest request) {
56 async.Future<GoogleLongrunningOperation> annotate(GoogleCloudVideointelligence V1beta1AnnotateVideoRequest request) {
57 var _url = null; 57 var _url = null;
58 var _queryParams = new core.Map(); 58 var _queryParams = new core.Map();
59 var _uploadMedia = null; 59 var _uploadMedia = null;
60 var _uploadOptions = null; 60 var _uploadOptions = null;
61 var _downloadOptions = commons.DownloadOptions.Metadata; 61 var _downloadOptions = commons.DownloadOptions.Metadata;
62 var _body = null; 62 var _body = null;
63 63
64 if (request != null) { 64 if (request != null) {
65 _body = convert.JSON.encode((request).toJson()); 65 _body = convert.JSON.encode((request).toJson());
66 } 66 }
67 67
68 _url = 'v1beta1/videos:annotate'; 68 _url = 'v1beta1/videos:annotate';
69 69
70 var _response = _requester.request(_url, 70 var _response = _requester.request(_url, "POST",
71 "POST", 71 body: _body,
72 body: _body, 72 queryParams: _queryParams,
73 queryParams: _queryParams, 73 uploadOptions: _uploadOptions,
74 uploadOptions: _uploadOptions, 74 uploadMedia: _uploadMedia,
75 uploadMedia: _uploadMedia, 75 downloadOptions: _downloadOptions);
76 downloadOptions: _downloadOptions); 76 return _response
77 return _response.then((data) => new GoogleLongrunningOperation.fromJson(data )); 77 .then((data) => new GoogleLongrunningOperation.fromJson(data));
78 } 78 }
79
80 } 79 }
81 80
82 81 /// Video annotation progress. Included in the `metadata`
83 82 /// field of the `Operation` returned by the `GetOperation`
84 /** 83 /// call of the `google::longrunning::Operations` service.
85 * Video annotation progress. Included in the `metadata`
86 * field of the `Operation` returned by the `GetOperation`
87 * call of the `google::longrunning::Operations` service.
88 */
89 class GoogleCloudVideointelligenceV1AnnotateVideoProgress { 84 class GoogleCloudVideointelligenceV1AnnotateVideoProgress {
90 /** Progress metadata for all videos specified in `AnnotateVideoRequest`. */ 85 /// Progress metadata for all videos specified in `AnnotateVideoRequest`.
91 core.List<GoogleCloudVideointelligenceV1VideoAnnotationProgress> annotationPro gress; 86 core.List<GoogleCloudVideointelligenceV1VideoAnnotationProgress>
87 annotationProgress;
92 88
93 GoogleCloudVideointelligenceV1AnnotateVideoProgress(); 89 GoogleCloudVideointelligenceV1AnnotateVideoProgress();
94 90
95 GoogleCloudVideointelligenceV1AnnotateVideoProgress.fromJson(core.Map _json) { 91 GoogleCloudVideointelligenceV1AnnotateVideoProgress.fromJson(core.Map _json) {
96 if (_json.containsKey("annotationProgress")) { 92 if (_json.containsKey("annotationProgress")) {
97 annotationProgress = _json["annotationProgress"].map((value) => new Google CloudVideointelligenceV1VideoAnnotationProgress.fromJson(value)).toList(); 93 annotationProgress = _json["annotationProgress"]
94 .map((value) =>
95 new GoogleCloudVideointelligenceV1VideoAnnotationProgress
96 .fromJson(value))
97 .toList();
98 } 98 }
99 } 99 }
100 100
101 core.Map<core.String, core.Object> toJson() { 101 core.Map<core.String, core.Object> toJson() {
102 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 102 final core.Map<core.String, core.Object> _json =
103 new core.Map<core.String, core.Object>();
103 if (annotationProgress != null) { 104 if (annotationProgress != null) {
104 _json["annotationProgress"] = annotationProgress.map((value) => (value).to Json()).toList(); 105 _json["annotationProgress"] =
106 annotationProgress.map((value) => (value).toJson()).toList();
105 } 107 }
106 return _json; 108 return _json;
107 } 109 }
108 } 110 }
109 111
110 /** 112 /// Video annotation response. Included in the `response`
111 * Video annotation response. Included in the `response` 113 /// field of the `Operation` returned by the `GetOperation`
112 * field of the `Operation` returned by the `GetOperation` 114 /// call of the `google::longrunning::Operations` service.
113 * call of the `google::longrunning::Operations` service.
114 */
115 class GoogleCloudVideointelligenceV1AnnotateVideoResponse { 115 class GoogleCloudVideointelligenceV1AnnotateVideoResponse {
116 /** Annotation results for all videos specified in `AnnotateVideoRequest`. */ 116 /// Annotation results for all videos specified in `AnnotateVideoRequest`.
117 core.List<GoogleCloudVideointelligenceV1VideoAnnotationResults> annotationResu lts; 117 core.List<GoogleCloudVideointelligenceV1VideoAnnotationResults>
118 annotationResults;
118 119
119 GoogleCloudVideointelligenceV1AnnotateVideoResponse(); 120 GoogleCloudVideointelligenceV1AnnotateVideoResponse();
120 121
121 GoogleCloudVideointelligenceV1AnnotateVideoResponse.fromJson(core.Map _json) { 122 GoogleCloudVideointelligenceV1AnnotateVideoResponse.fromJson(core.Map _json) {
122 if (_json.containsKey("annotationResults")) { 123 if (_json.containsKey("annotationResults")) {
123 annotationResults = _json["annotationResults"].map((value) => new GoogleCl oudVideointelligenceV1VideoAnnotationResults.fromJson(value)).toList(); 124 annotationResults = _json["annotationResults"]
125 .map((value) =>
126 new GoogleCloudVideointelligenceV1VideoAnnotationResults.fromJson(
127 value))
128 .toList();
124 } 129 }
125 } 130 }
126 131
127 core.Map<core.String, core.Object> toJson() { 132 core.Map<core.String, core.Object> toJson() {
128 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 133 final core.Map<core.String, core.Object> _json =
134 new core.Map<core.String, core.Object>();
129 if (annotationResults != null) { 135 if (annotationResults != null) {
130 _json["annotationResults"] = annotationResults.map((value) => (value).toJs on()).toList(); 136 _json["annotationResults"] =
137 annotationResults.map((value) => (value).toJson()).toList();
131 } 138 }
132 return _json; 139 return _json;
133 } 140 }
134 } 141 }
135 142
136 /** Label annotation. */ 143 /// Label annotation.
137 class GoogleCloudVideointelligenceV1LabelAnnotation { 144 class GoogleCloudVideointelligenceV1LabelAnnotation {
138 /** Textual description, e.g. `Fixed-gear bicycle`. */ 145 /// Textual description, e.g. `Fixed-gear bicycle`.
139 core.String description; 146 core.String description;
140 /** Language code for `description` in BCP-47 format. */ 147
148 /// Language code for `description` in BCP-47 format.
141 core.String languageCode; 149 core.String languageCode;
142 /** Where the label was detected and with what confidence. */ 150
151 /// Where the label was detected and with what confidence.
143 core.List<GoogleCloudVideointelligenceV1LabelLocation> locations; 152 core.List<GoogleCloudVideointelligenceV1LabelLocation> locations;
144 153
145 GoogleCloudVideointelligenceV1LabelAnnotation(); 154 GoogleCloudVideointelligenceV1LabelAnnotation();
146 155
147 GoogleCloudVideointelligenceV1LabelAnnotation.fromJson(core.Map _json) { 156 GoogleCloudVideointelligenceV1LabelAnnotation.fromJson(core.Map _json) {
148 if (_json.containsKey("description")) { 157 if (_json.containsKey("description")) {
149 description = _json["description"]; 158 description = _json["description"];
150 } 159 }
151 if (_json.containsKey("languageCode")) { 160 if (_json.containsKey("languageCode")) {
152 languageCode = _json["languageCode"]; 161 languageCode = _json["languageCode"];
153 } 162 }
154 if (_json.containsKey("locations")) { 163 if (_json.containsKey("locations")) {
155 locations = _json["locations"].map((value) => new GoogleCloudVideointellig enceV1LabelLocation.fromJson(value)).toList(); 164 locations = _json["locations"]
165 .map((value) =>
166 new GoogleCloudVideointelligenceV1LabelLocation.fromJson(value))
167 .toList();
156 } 168 }
157 } 169 }
158 170
159 core.Map<core.String, core.Object> toJson() { 171 core.Map<core.String, core.Object> toJson() {
160 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 172 final core.Map<core.String, core.Object> _json =
173 new core.Map<core.String, core.Object>();
161 if (description != null) { 174 if (description != null) {
162 _json["description"] = description; 175 _json["description"] = description;
163 } 176 }
164 if (languageCode != null) { 177 if (languageCode != null) {
165 _json["languageCode"] = languageCode; 178 _json["languageCode"] = languageCode;
166 } 179 }
167 if (locations != null) { 180 if (locations != null) {
168 _json["locations"] = locations.map((value) => (value).toJson()).toList(); 181 _json["locations"] = locations.map((value) => (value).toJson()).toList();
169 } 182 }
170 return _json; 183 return _json;
171 } 184 }
172 } 185 }
173 186
174 /** Label location. */ 187 /// Label location.
175 class GoogleCloudVideointelligenceV1LabelLocation { 188 class GoogleCloudVideointelligenceV1LabelLocation {
176 /** Confidence that the label is accurate. Range: [0, 1]. */ 189 /// Confidence that the label is accurate. Range: [0, 1].
177 core.double confidence; 190 core.double confidence;
178 /** 191
179 * Label level. 192 /// Label level.
180 * Possible string values are: 193 /// Possible string values are:
181 * - "LABEL_LEVEL_UNSPECIFIED" : Unspecified. 194 /// - "LABEL_LEVEL_UNSPECIFIED" : Unspecified.
182 * - "VIDEO_LEVEL" : Video-level. Corresponds to the whole video. 195 /// - "VIDEO_LEVEL" : Video-level. Corresponds to the whole video.
183 * - "SEGMENT_LEVEL" : Segment-level. Corresponds to one of 196 /// - "SEGMENT_LEVEL" : Segment-level. Corresponds to one of
184 * `AnnotateSpec.segments`. 197 /// `AnnotateSpec.segments`.
185 * - "SHOT_LEVEL" : Shot-level. Corresponds to a single shot (i.e. a series of 198 /// - "SHOT_LEVEL" : Shot-level. Corresponds to a single shot (i.e. a series
186 * frames 199 /// of frames
187 * without a major camera position or background change). 200 /// without a major camera position or background change).
188 * - "FRAME_LEVEL" : Frame-level. Corresponds to a single video frame. 201 /// - "FRAME_LEVEL" : Frame-level. Corresponds to a single video frame.
189 */
190 core.String level; 202 core.String level;
191 /** 203
192 * Video segment. Unset for video-level labels. 204 /// Video segment. Unset for video-level labels.
193 * Set to a frame timestamp for frame-level labels. 205 /// Set to a frame timestamp for frame-level labels.
194 * Otherwise, corresponds to one of `AnnotateSpec.segments` 206 /// Otherwise, corresponds to one of `AnnotateSpec.segments`
195 * (if specified) or to shot boundaries (if requested). 207 /// (if specified) or to shot boundaries (if requested).
196 */
197 GoogleCloudVideointelligenceV1VideoSegment segment; 208 GoogleCloudVideointelligenceV1VideoSegment segment;
198 209
199 GoogleCloudVideointelligenceV1LabelLocation(); 210 GoogleCloudVideointelligenceV1LabelLocation();
200 211
201 GoogleCloudVideointelligenceV1LabelLocation.fromJson(core.Map _json) { 212 GoogleCloudVideointelligenceV1LabelLocation.fromJson(core.Map _json) {
202 if (_json.containsKey("confidence")) { 213 if (_json.containsKey("confidence")) {
203 confidence = _json["confidence"]; 214 confidence = _json["confidence"];
204 } 215 }
205 if (_json.containsKey("level")) { 216 if (_json.containsKey("level")) {
206 level = _json["level"]; 217 level = _json["level"];
207 } 218 }
208 if (_json.containsKey("segment")) { 219 if (_json.containsKey("segment")) {
209 segment = new GoogleCloudVideointelligenceV1VideoSegment.fromJson(_json["s egment"]); 220 segment = new GoogleCloudVideointelligenceV1VideoSegment.fromJson(
221 _json["segment"]);
210 } 222 }
211 } 223 }
212 224
213 core.Map<core.String, core.Object> toJson() { 225 core.Map<core.String, core.Object> toJson() {
214 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 226 final core.Map<core.String, core.Object> _json =
227 new core.Map<core.String, core.Object>();
215 if (confidence != null) { 228 if (confidence != null) {
216 _json["confidence"] = confidence; 229 _json["confidence"] = confidence;
217 } 230 }
218 if (level != null) { 231 if (level != null) {
219 _json["level"] = level; 232 _json["level"] = level;
220 } 233 }
221 if (segment != null) { 234 if (segment != null) {
222 _json["segment"] = (segment).toJson(); 235 _json["segment"] = (segment).toJson();
223 } 236 }
224 return _json; 237 return _json;
225 } 238 }
226 } 239 }
227 240
228 /** 241 /// Safe search annotation (based on per-frame visual signals only).
229 * Safe search annotation (based on per-frame visual signals only). 242 /// If no unsafe content has been detected in a frame, no annotations
230 * If no unsafe content has been detected in a frame, no annotations 243 /// are present for that frame.
231 * are present for that frame.
232 */
233 class GoogleCloudVideointelligenceV1SafeSearchAnnotation { 244 class GoogleCloudVideointelligenceV1SafeSearchAnnotation {
234 /** 245 /// Likelihood of adult content.
235 * Likelihood of adult content. 246 /// Possible string values are:
236 * Possible string values are: 247 /// - "UNKNOWN" : Unknown likelihood.
237 * - "UNKNOWN" : Unknown likelihood. 248 /// - "VERY_UNLIKELY" : Very unlikely.
238 * - "VERY_UNLIKELY" : Very unlikely. 249 /// - "UNLIKELY" : Unlikely.
239 * - "UNLIKELY" : Unlikely. 250 /// - "POSSIBLE" : Possible.
240 * - "POSSIBLE" : Possible. 251 /// - "LIKELY" : Likely.
241 * - "LIKELY" : Likely. 252 /// - "VERY_LIKELY" : Very likely.
242 * - "VERY_LIKELY" : Very likely.
243 */
244 core.String adult; 253 core.String adult;
245 /** 254
246 * Time-offset, relative to the beginning of the video, 255 /// Time-offset, relative to the beginning of the video,
247 * corresponding to the video frame for this annotation. 256 /// corresponding to the video frame for this annotation.
248 */
249 core.String time; 257 core.String time;
250 258
251 GoogleCloudVideointelligenceV1SafeSearchAnnotation(); 259 GoogleCloudVideointelligenceV1SafeSearchAnnotation();
252 260
253 GoogleCloudVideointelligenceV1SafeSearchAnnotation.fromJson(core.Map _json) { 261 GoogleCloudVideointelligenceV1SafeSearchAnnotation.fromJson(core.Map _json) {
254 if (_json.containsKey("adult")) { 262 if (_json.containsKey("adult")) {
255 adult = _json["adult"]; 263 adult = _json["adult"];
256 } 264 }
257 if (_json.containsKey("time")) { 265 if (_json.containsKey("time")) {
258 time = _json["time"]; 266 time = _json["time"];
259 } 267 }
260 } 268 }
261 269
262 core.Map<core.String, core.Object> toJson() { 270 core.Map<core.String, core.Object> toJson() {
263 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 271 final core.Map<core.String, core.Object> _json =
272 new core.Map<core.String, core.Object>();
264 if (adult != null) { 273 if (adult != null) {
265 _json["adult"] = adult; 274 _json["adult"] = adult;
266 } 275 }
267 if (time != null) { 276 if (time != null) {
268 _json["time"] = time; 277 _json["time"] = time;
269 } 278 }
270 return _json; 279 return _json;
271 } 280 }
272 } 281 }
273 282
274 /** Annotation progress for a single video. */ 283 /// Annotation progress for a single video.
275 class GoogleCloudVideointelligenceV1VideoAnnotationProgress { 284 class GoogleCloudVideointelligenceV1VideoAnnotationProgress {
276 /** 285 /// Video file location in
277 * Video file location in 286 /// [Google Cloud Storage](https://cloud.google.com/storage/).
278 * [Google Cloud Storage](https://cloud.google.com/storage/).
279 */
280 core.String inputUri; 287 core.String inputUri;
281 /** 288
282 * Approximate percentage processed thus far. 289 /// Approximate percentage processed thus far.
283 * Guaranteed to be 100 when fully processed. 290 /// Guaranteed to be 100 when fully processed.
284 */
285 core.int progressPercent; 291 core.int progressPercent;
286 /** Time when the request was received. */ 292
293 /// Time when the request was received.
287 core.String startTime; 294 core.String startTime;
288 /** Time of the most recent update. */ 295
296 /// Time of the most recent update.
289 core.String updateTime; 297 core.String updateTime;
290 298
291 GoogleCloudVideointelligenceV1VideoAnnotationProgress(); 299 GoogleCloudVideointelligenceV1VideoAnnotationProgress();
292 300
293 GoogleCloudVideointelligenceV1VideoAnnotationProgress.fromJson(core.Map _json) { 301 GoogleCloudVideointelligenceV1VideoAnnotationProgress.fromJson(
302 core.Map _json) {
294 if (_json.containsKey("inputUri")) { 303 if (_json.containsKey("inputUri")) {
295 inputUri = _json["inputUri"]; 304 inputUri = _json["inputUri"];
296 } 305 }
297 if (_json.containsKey("progressPercent")) { 306 if (_json.containsKey("progressPercent")) {
298 progressPercent = _json["progressPercent"]; 307 progressPercent = _json["progressPercent"];
299 } 308 }
300 if (_json.containsKey("startTime")) { 309 if (_json.containsKey("startTime")) {
301 startTime = _json["startTime"]; 310 startTime = _json["startTime"];
302 } 311 }
303 if (_json.containsKey("updateTime")) { 312 if (_json.containsKey("updateTime")) {
304 updateTime = _json["updateTime"]; 313 updateTime = _json["updateTime"];
305 } 314 }
306 } 315 }
307 316
308 core.Map<core.String, core.Object> toJson() { 317 core.Map<core.String, core.Object> toJson() {
309 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 318 final core.Map<core.String, core.Object> _json =
319 new core.Map<core.String, core.Object>();
310 if (inputUri != null) { 320 if (inputUri != null) {
311 _json["inputUri"] = inputUri; 321 _json["inputUri"] = inputUri;
312 } 322 }
313 if (progressPercent != null) { 323 if (progressPercent != null) {
314 _json["progressPercent"] = progressPercent; 324 _json["progressPercent"] = progressPercent;
315 } 325 }
316 if (startTime != null) { 326 if (startTime != null) {
317 _json["startTime"] = startTime; 327 _json["startTime"] = startTime;
318 } 328 }
319 if (updateTime != null) { 329 if (updateTime != null) {
320 _json["updateTime"] = updateTime; 330 _json["updateTime"] = updateTime;
321 } 331 }
322 return _json; 332 return _json;
323 } 333 }
324 } 334 }
325 335
326 /** Annotation results for a single video. */ 336 /// Annotation results for a single video.
327 class GoogleCloudVideointelligenceV1VideoAnnotationResults { 337 class GoogleCloudVideointelligenceV1VideoAnnotationResults {
328 /** 338 /// If set, indicates an error. Note that for a single `AnnotateVideoRequest`
329 * If set, indicates an error. Note that for a single `AnnotateVideoRequest` 339 /// some videos may succeed and some may fail.
330 * some videos may succeed and some may fail.
331 */
332 GoogleRpcStatus error; 340 GoogleRpcStatus error;
333 /** 341
334 * Video file location in 342 /// Video file location in
335 * [Google Cloud Storage](https://cloud.google.com/storage/). 343 /// [Google Cloud Storage](https://cloud.google.com/storage/).
336 */
337 core.String inputUri; 344 core.String inputUri;
338 /** Label annotations. There is exactly one element for each unique label. */ 345
346 /// Label annotations. There is exactly one element for each unique label.
339 core.List<GoogleCloudVideointelligenceV1LabelAnnotation> labelAnnotations; 347 core.List<GoogleCloudVideointelligenceV1LabelAnnotation> labelAnnotations;
340 /** Safe search annotations. */ 348
341 core.List<GoogleCloudVideointelligenceV1SafeSearchAnnotation> safeSearchAnnota tions; 349 /// Safe search annotations.
342 /** Shot annotations. Each shot is represented as a video segment. */ 350 core.List<GoogleCloudVideointelligenceV1SafeSearchAnnotation>
351 safeSearchAnnotations;
352
353 /// Shot annotations. Each shot is represented as a video segment.
343 core.List<GoogleCloudVideointelligenceV1VideoSegment> shotAnnotations; 354 core.List<GoogleCloudVideointelligenceV1VideoSegment> shotAnnotations;
344 355
345 GoogleCloudVideointelligenceV1VideoAnnotationResults(); 356 GoogleCloudVideointelligenceV1VideoAnnotationResults();
346 357
347 GoogleCloudVideointelligenceV1VideoAnnotationResults.fromJson(core.Map _json) { 358 GoogleCloudVideointelligenceV1VideoAnnotationResults.fromJson(
359 core.Map _json) {
348 if (_json.containsKey("error")) { 360 if (_json.containsKey("error")) {
349 error = new GoogleRpcStatus.fromJson(_json["error"]); 361 error = new GoogleRpcStatus.fromJson(_json["error"]);
350 } 362 }
351 if (_json.containsKey("inputUri")) { 363 if (_json.containsKey("inputUri")) {
352 inputUri = _json["inputUri"]; 364 inputUri = _json["inputUri"];
353 } 365 }
354 if (_json.containsKey("labelAnnotations")) { 366 if (_json.containsKey("labelAnnotations")) {
355 labelAnnotations = _json["labelAnnotations"].map((value) => new GoogleClou dVideointelligenceV1LabelAnnotation.fromJson(value)).toList(); 367 labelAnnotations = _json["labelAnnotations"]
368 .map((value) =>
369 new GoogleCloudVideointelligenceV1LabelAnnotation.fromJson(value))
370 .toList();
356 } 371 }
357 if (_json.containsKey("safeSearchAnnotations")) { 372 if (_json.containsKey("safeSearchAnnotations")) {
358 safeSearchAnnotations = _json["safeSearchAnnotations"].map((value) => new GoogleCloudVideointelligenceV1SafeSearchAnnotation.fromJson(value)).toList(); 373 safeSearchAnnotations = _json["safeSearchAnnotations"]
374 .map((value) =>
375 new GoogleCloudVideointelligenceV1SafeSearchAnnotation.fromJson(
376 value))
377 .toList();
359 } 378 }
360 if (_json.containsKey("shotAnnotations")) { 379 if (_json.containsKey("shotAnnotations")) {
361 shotAnnotations = _json["shotAnnotations"].map((value) => new GoogleCloudV ideointelligenceV1VideoSegment.fromJson(value)).toList(); 380 shotAnnotations = _json["shotAnnotations"]
381 .map((value) =>
382 new GoogleCloudVideointelligenceV1VideoSegment.fromJson(value))
383 .toList();
362 } 384 }
363 } 385 }
364 386
365 core.Map<core.String, core.Object> toJson() { 387 core.Map<core.String, core.Object> toJson() {
366 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 388 final core.Map<core.String, core.Object> _json =
389 new core.Map<core.String, core.Object>();
367 if (error != null) { 390 if (error != null) {
368 _json["error"] = (error).toJson(); 391 _json["error"] = (error).toJson();
369 } 392 }
370 if (inputUri != null) { 393 if (inputUri != null) {
371 _json["inputUri"] = inputUri; 394 _json["inputUri"] = inputUri;
372 } 395 }
373 if (labelAnnotations != null) { 396 if (labelAnnotations != null) {
374 _json["labelAnnotations"] = labelAnnotations.map((value) => (value).toJson ()).toList(); 397 _json["labelAnnotations"] =
398 labelAnnotations.map((value) => (value).toJson()).toList();
375 } 399 }
376 if (safeSearchAnnotations != null) { 400 if (safeSearchAnnotations != null) {
377 _json["safeSearchAnnotations"] = safeSearchAnnotations.map((value) => (val ue).toJson()).toList(); 401 _json["safeSearchAnnotations"] =
402 safeSearchAnnotations.map((value) => (value).toJson()).toList();
378 } 403 }
379 if (shotAnnotations != null) { 404 if (shotAnnotations != null) {
380 _json["shotAnnotations"] = shotAnnotations.map((value) => (value).toJson() ).toList(); 405 _json["shotAnnotations"] =
406 shotAnnotations.map((value) => (value).toJson()).toList();
381 } 407 }
382 return _json; 408 return _json;
383 } 409 }
384 } 410 }
385 411
386 /** Video segment. */ 412 /// Video segment.
387 class GoogleCloudVideointelligenceV1VideoSegment { 413 class GoogleCloudVideointelligenceV1VideoSegment {
388 /** 414 /// Time-offset, relative to the beginning of the video,
389 * Time-offset, relative to the beginning of the video, 415 /// corresponding to the end of the segment (inclusive).
390 * corresponding to the end of the segment (inclusive).
391 */
392 core.String endTime; 416 core.String endTime;
393 /** 417
394 * Time-offset, relative to the beginning of the video, 418 /// Time-offset, relative to the beginning of the video,
395 * corresponding to the start of the segment (inclusive). 419 /// corresponding to the start of the segment (inclusive).
396 */
397 core.String startTime; 420 core.String startTime;
398 421
399 GoogleCloudVideointelligenceV1VideoSegment(); 422 GoogleCloudVideointelligenceV1VideoSegment();
400 423
401 GoogleCloudVideointelligenceV1VideoSegment.fromJson(core.Map _json) { 424 GoogleCloudVideointelligenceV1VideoSegment.fromJson(core.Map _json) {
402 if (_json.containsKey("endTime")) { 425 if (_json.containsKey("endTime")) {
403 endTime = _json["endTime"]; 426 endTime = _json["endTime"];
404 } 427 }
405 if (_json.containsKey("startTime")) { 428 if (_json.containsKey("startTime")) {
406 startTime = _json["startTime"]; 429 startTime = _json["startTime"];
407 } 430 }
408 } 431 }
409 432
410 core.Map<core.String, core.Object> toJson() { 433 core.Map<core.String, core.Object> toJson() {
411 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 434 final core.Map<core.String, core.Object> _json =
435 new core.Map<core.String, core.Object>();
412 if (endTime != null) { 436 if (endTime != null) {
413 _json["endTime"] = endTime; 437 _json["endTime"] = endTime;
414 } 438 }
415 if (startTime != null) { 439 if (startTime != null) {
416 _json["startTime"] = startTime; 440 _json["startTime"] = startTime;
417 } 441 }
418 return _json; 442 return _json;
419 } 443 }
420 } 444 }
421 445
422 /** 446 /// Video annotation progress. Included in the `metadata`
423 * Video annotation progress. Included in the `metadata` 447 /// field of the `Operation` returned by the `GetOperation`
424 * field of the `Operation` returned by the `GetOperation` 448 /// call of the `google::longrunning::Operations` service.
425 * call of the `google::longrunning::Operations` service.
426 */
427 class GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress { 449 class GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress {
428 /** Progress metadata for all videos specified in `AnnotateVideoRequest`. */ 450 /// Progress metadata for all videos specified in `AnnotateVideoRequest`.
429 core.List<GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress> annotati onProgress; 451 core.List<GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress>
452 annotationProgress;
430 453
431 GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress(); 454 GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress();
432 455
433 GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress.fromJson(core.Map _js on) { 456 GoogleCloudVideointelligenceV1beta1AnnotateVideoProgress.fromJson(
457 core.Map _json) {
434 if (_json.containsKey("annotationProgress")) { 458 if (_json.containsKey("annotationProgress")) {
435 annotationProgress = _json["annotationProgress"].map((value) => new Google CloudVideointelligenceV1beta1VideoAnnotationProgress.fromJson(value)).toList(); 459 annotationProgress = _json["annotationProgress"]
460 .map((value) =>
461 new GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress
462 .fromJson(value))
463 .toList();
436 } 464 }
437 } 465 }
438 466
439 core.Map<core.String, core.Object> toJson() { 467 core.Map<core.String, core.Object> toJson() {
440 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 468 final core.Map<core.String, core.Object> _json =
469 new core.Map<core.String, core.Object>();
441 if (annotationProgress != null) { 470 if (annotationProgress != null) {
442 _json["annotationProgress"] = annotationProgress.map((value) => (value).to Json()).toList(); 471 _json["annotationProgress"] =
472 annotationProgress.map((value) => (value).toJson()).toList();
443 } 473 }
444 return _json; 474 return _json;
445 } 475 }
446 } 476 }
447 477
448 /** Video annotation request. */ 478 /// Video annotation request.
449 class GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest { 479 class GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest {
450 /** Requested video annotation features. */ 480 /// Requested video annotation features.
451 core.List<core.String> features; 481 core.List<core.String> features;
452 /** 482
453 * The video data bytes. Encoding: base64. If unset, the input video(s) 483 /// The video data bytes. Encoding: base64. If unset, the input video(s)
454 * should be specified via `input_uri`. If set, `input_uri` should be unset. 484 /// should be specified via `input_uri`. If set, `input_uri` should be unset.
455 */
456 core.String inputContent; 485 core.String inputContent;
457 /** 486
458 * Input video location. Currently, only 487 /// Input video location. Currently, only
459 * [Google Cloud Storage](https://cloud.google.com/storage/) URIs are 488 /// [Google Cloud Storage](https://cloud.google.com/storage/) URIs are
460 * supported, which must be specified in the following format: 489 /// supported, which must be specified in the following format:
461 * `gs://bucket-id/object-id` (other URI formats return 490 /// `gs://bucket-id/object-id` (other URI formats return
462 * google.rpc.Code.INVALID_ARGUMENT). For more information, see 491 /// google.rpc.Code.INVALID_ARGUMENT). For more information, see
463 * [Request URIs](/storage/docs/reference-uris). 492 /// [Request URIs](/storage/docs/reference-uris).
464 * A video URI may include wildcards in `object-id`, and thus identify 493 /// A video URI may include wildcards in `object-id`, and thus identify
465 * multiple videos. Supported wildcards: '*' to match 0 or more characters; 494 /// multiple videos. Supported wildcards: '*' to match 0 or more characters;
466 * '?' to match 1 character. If unset, the input video should be embedded 495 /// '?' to match 1 character. If unset, the input video should be embedded
467 * in the request as `input_content`. If set, `input_content` should be unset. 496 /// in the request as `input_content`. If set, `input_content` should be
468 */ 497 /// unset.
469 core.String inputUri; 498 core.String inputUri;
470 /** 499
471 * Optional cloud region where annotation should take place. Supported cloud 500 /// Optional cloud region where annotation should take place. Supported cloud
472 * regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region 501 /// regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no
473 * is specified, a region will be determined based on video file location. 502 /// region
474 */ 503 /// is specified, a region will be determined based on video file location.
475 core.String locationId; 504 core.String locationId;
476 /** 505
477 * Optional location where the output (in JSON format) should be stored. 506 /// Optional location where the output (in JSON format) should be stored.
478 * Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) 507 /// Currently, only [Google Cloud Storage](https://cloud.google.com/storage/)
479 * URIs are supported, which must be specified in the following format: 508 /// URIs are supported, which must be specified in the following format:
480 * `gs://bucket-id/object-id` (other URI formats return 509 /// `gs://bucket-id/object-id` (other URI formats return
481 * google.rpc.Code.INVALID_ARGUMENT). For more information, see 510 /// google.rpc.Code.INVALID_ARGUMENT). For more information, see
482 * [Request URIs](/storage/docs/reference-uris). 511 /// [Request URIs](/storage/docs/reference-uris).
483 */
484 core.String outputUri; 512 core.String outputUri;
485 /** Additional video context and/or feature-specific parameters. */ 513
514 /// Additional video context and/or feature-specific parameters.
486 GoogleCloudVideointelligenceV1beta1VideoContext videoContext; 515 GoogleCloudVideointelligenceV1beta1VideoContext videoContext;
487 516
488 GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest(); 517 GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest();
489 518
490 GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest.fromJson(core.Map _jso n) { 519 GoogleCloudVideointelligenceV1beta1AnnotateVideoRequest.fromJson(
520 core.Map _json) {
491 if (_json.containsKey("features")) { 521 if (_json.containsKey("features")) {
492 features = _json["features"]; 522 features = _json["features"];
493 } 523 }
494 if (_json.containsKey("inputContent")) { 524 if (_json.containsKey("inputContent")) {
495 inputContent = _json["inputContent"]; 525 inputContent = _json["inputContent"];
496 } 526 }
497 if (_json.containsKey("inputUri")) { 527 if (_json.containsKey("inputUri")) {
498 inputUri = _json["inputUri"]; 528 inputUri = _json["inputUri"];
499 } 529 }
500 if (_json.containsKey("locationId")) { 530 if (_json.containsKey("locationId")) {
501 locationId = _json["locationId"]; 531 locationId = _json["locationId"];
502 } 532 }
503 if (_json.containsKey("outputUri")) { 533 if (_json.containsKey("outputUri")) {
504 outputUri = _json["outputUri"]; 534 outputUri = _json["outputUri"];
505 } 535 }
506 if (_json.containsKey("videoContext")) { 536 if (_json.containsKey("videoContext")) {
507 videoContext = new GoogleCloudVideointelligenceV1beta1VideoContext.fromJso n(_json["videoContext"]); 537 videoContext =
538 new GoogleCloudVideointelligenceV1beta1VideoContext.fromJson(
539 _json["videoContext"]);
508 } 540 }
509 } 541 }
510 542
511 core.Map<core.String, core.Object> toJson() { 543 core.Map<core.String, core.Object> toJson() {
512 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 544 final core.Map<core.String, core.Object> _json =
545 new core.Map<core.String, core.Object>();
513 if (features != null) { 546 if (features != null) {
514 _json["features"] = features; 547 _json["features"] = features;
515 } 548 }
516 if (inputContent != null) { 549 if (inputContent != null) {
517 _json["inputContent"] = inputContent; 550 _json["inputContent"] = inputContent;
518 } 551 }
519 if (inputUri != null) { 552 if (inputUri != null) {
520 _json["inputUri"] = inputUri; 553 _json["inputUri"] = inputUri;
521 } 554 }
522 if (locationId != null) { 555 if (locationId != null) {
523 _json["locationId"] = locationId; 556 _json["locationId"] = locationId;
524 } 557 }
525 if (outputUri != null) { 558 if (outputUri != null) {
526 _json["outputUri"] = outputUri; 559 _json["outputUri"] = outputUri;
527 } 560 }
528 if (videoContext != null) { 561 if (videoContext != null) {
529 _json["videoContext"] = (videoContext).toJson(); 562 _json["videoContext"] = (videoContext).toJson();
530 } 563 }
531 return _json; 564 return _json;
532 } 565 }
533 } 566 }
534 567
535 /** 568 /// Video annotation response. Included in the `response`
536 * Video annotation response. Included in the `response` 569 /// field of the `Operation` returned by the `GetOperation`
537 * field of the `Operation` returned by the `GetOperation` 570 /// call of the `google::longrunning::Operations` service.
538 * call of the `google::longrunning::Operations` service.
539 */
540 class GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse { 571 class GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse {
541 /** Annotation results for all videos specified in `AnnotateVideoRequest`. */ 572 /// Annotation results for all videos specified in `AnnotateVideoRequest`.
542 core.List<GoogleCloudVideointelligenceV1beta1VideoAnnotationResults> annotatio nResults; 573 core.List<GoogleCloudVideointelligenceV1beta1VideoAnnotationResults>
574 annotationResults;
543 575
544 GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse(); 576 GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse();
545 577
546 GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse.fromJson(core.Map _js on) { 578 GoogleCloudVideointelligenceV1beta1AnnotateVideoResponse.fromJson(
579 core.Map _json) {
547 if (_json.containsKey("annotationResults")) { 580 if (_json.containsKey("annotationResults")) {
548 annotationResults = _json["annotationResults"].map((value) => new GoogleCl oudVideointelligenceV1beta1VideoAnnotationResults.fromJson(value)).toList(); 581 annotationResults = _json["annotationResults"]
582 .map((value) =>
583 new GoogleCloudVideointelligenceV1beta1VideoAnnotationResults
584 .fromJson(value))
585 .toList();
549 } 586 }
550 } 587 }
551 588
552 core.Map<core.String, core.Object> toJson() { 589 core.Map<core.String, core.Object> toJson() {
553 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 590 final core.Map<core.String, core.Object> _json =
591 new core.Map<core.String, core.Object>();
554 if (annotationResults != null) { 592 if (annotationResults != null) {
555 _json["annotationResults"] = annotationResults.map((value) => (value).toJs on()).toList(); 593 _json["annotationResults"] =
594 annotationResults.map((value) => (value).toJson()).toList();
556 } 595 }
557 return _json; 596 return _json;
558 } 597 }
559 } 598 }
560 599
561 /** Label annotation. */ 600 /// Label annotation.
562 class GoogleCloudVideointelligenceV1beta1LabelAnnotation { 601 class GoogleCloudVideointelligenceV1beta1LabelAnnotation {
563 /** Textual description, e.g. `Fixed-gear bicycle`. */ 602 /// Textual description, e.g. `Fixed-gear bicycle`.
564 core.String description; 603 core.String description;
565 /** Language code for `description` in BCP-47 format. */ 604
605 /// Language code for `description` in BCP-47 format.
566 core.String languageCode; 606 core.String languageCode;
567 /** Where the label was detected and with what confidence. */ 607
608 /// Where the label was detected and with what confidence.
568 core.List<GoogleCloudVideointelligenceV1beta1LabelLocation> locations; 609 core.List<GoogleCloudVideointelligenceV1beta1LabelLocation> locations;
569 610
570 GoogleCloudVideointelligenceV1beta1LabelAnnotation(); 611 GoogleCloudVideointelligenceV1beta1LabelAnnotation();
571 612
572 GoogleCloudVideointelligenceV1beta1LabelAnnotation.fromJson(core.Map _json) { 613 GoogleCloudVideointelligenceV1beta1LabelAnnotation.fromJson(core.Map _json) {
573 if (_json.containsKey("description")) { 614 if (_json.containsKey("description")) {
574 description = _json["description"]; 615 description = _json["description"];
575 } 616 }
576 if (_json.containsKey("languageCode")) { 617 if (_json.containsKey("languageCode")) {
577 languageCode = _json["languageCode"]; 618 languageCode = _json["languageCode"];
578 } 619 }
579 if (_json.containsKey("locations")) { 620 if (_json.containsKey("locations")) {
580 locations = _json["locations"].map((value) => new GoogleCloudVideointellig enceV1beta1LabelLocation.fromJson(value)).toList(); 621 locations = _json["locations"]
622 .map((value) =>
623 new GoogleCloudVideointelligenceV1beta1LabelLocation.fromJson(
624 value))
625 .toList();
581 } 626 }
582 } 627 }
583 628
584 core.Map<core.String, core.Object> toJson() { 629 core.Map<core.String, core.Object> toJson() {
585 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 630 final core.Map<core.String, core.Object> _json =
631 new core.Map<core.String, core.Object>();
586 if (description != null) { 632 if (description != null) {
587 _json["description"] = description; 633 _json["description"] = description;
588 } 634 }
589 if (languageCode != null) { 635 if (languageCode != null) {
590 _json["languageCode"] = languageCode; 636 _json["languageCode"] = languageCode;
591 } 637 }
592 if (locations != null) { 638 if (locations != null) {
593 _json["locations"] = locations.map((value) => (value).toJson()).toList(); 639 _json["locations"] = locations.map((value) => (value).toJson()).toList();
594 } 640 }
595 return _json; 641 return _json;
596 } 642 }
597 } 643 }
598 644
599 /** Label location. */ 645 /// Label location.
600 class GoogleCloudVideointelligenceV1beta1LabelLocation { 646 class GoogleCloudVideointelligenceV1beta1LabelLocation {
601 /** Confidence that the label is accurate. Range: [0, 1]. */ 647 /// Confidence that the label is accurate. Range: [0, 1].
602 core.double confidence; 648 core.double confidence;
603 /** 649
604 * Label level. 650 /// Label level.
605 * Possible string values are: 651 /// Possible string values are:
606 * - "LABEL_LEVEL_UNSPECIFIED" : Unspecified. 652 /// - "LABEL_LEVEL_UNSPECIFIED" : Unspecified.
607 * - "VIDEO_LEVEL" : Video-level. Corresponds to the whole video. 653 /// - "VIDEO_LEVEL" : Video-level. Corresponds to the whole video.
608 * - "SEGMENT_LEVEL" : Segment-level. Corresponds to one of 654 /// - "SEGMENT_LEVEL" : Segment-level. Corresponds to one of
609 * `AnnotateSpec.segments`. 655 /// `AnnotateSpec.segments`.
610 * - "SHOT_LEVEL" : Shot-level. Corresponds to a single shot (i.e. a series of 656 /// - "SHOT_LEVEL" : Shot-level. Corresponds to a single shot (i.e. a series
611 * frames 657 /// of frames
612 * without a major camera position or background change). 658 /// without a major camera position or background change).
613 * - "FRAME_LEVEL" : Frame-level. Corresponds to a single video frame. 659 /// - "FRAME_LEVEL" : Frame-level. Corresponds to a single video frame.
614 */
615 core.String level; 660 core.String level;
616 /** 661
617 * Video segment. Set to [-1, -1] for video-level labels. 662 /// Video segment. Set to [-1, -1] for video-level labels.
618 * Set to [timestamp, timestamp] for frame-level labels. 663 /// Set to [timestamp, timestamp] for frame-level labels.
619 * Otherwise, corresponds to one of `AnnotateSpec.segments` 664 /// Otherwise, corresponds to one of `AnnotateSpec.segments`
620 * (if specified) or to shot boundaries (if requested). 665 /// (if specified) or to shot boundaries (if requested).
621 */
622 GoogleCloudVideointelligenceV1beta1VideoSegment segment; 666 GoogleCloudVideointelligenceV1beta1VideoSegment segment;
623 667
624 GoogleCloudVideointelligenceV1beta1LabelLocation(); 668 GoogleCloudVideointelligenceV1beta1LabelLocation();
625 669
626 GoogleCloudVideointelligenceV1beta1LabelLocation.fromJson(core.Map _json) { 670 GoogleCloudVideointelligenceV1beta1LabelLocation.fromJson(core.Map _json) {
627 if (_json.containsKey("confidence")) { 671 if (_json.containsKey("confidence")) {
628 confidence = _json["confidence"]; 672 confidence = _json["confidence"];
629 } 673 }
630 if (_json.containsKey("level")) { 674 if (_json.containsKey("level")) {
631 level = _json["level"]; 675 level = _json["level"];
632 } 676 }
633 if (_json.containsKey("segment")) { 677 if (_json.containsKey("segment")) {
634 segment = new GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(_js on["segment"]); 678 segment = new GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(
679 _json["segment"]);
635 } 680 }
636 } 681 }
637 682
638 core.Map<core.String, core.Object> toJson() { 683 core.Map<core.String, core.Object> toJson() {
639 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 684 final core.Map<core.String, core.Object> _json =
685 new core.Map<core.String, core.Object>();
640 if (confidence != null) { 686 if (confidence != null) {
641 _json["confidence"] = confidence; 687 _json["confidence"] = confidence;
642 } 688 }
643 if (level != null) { 689 if (level != null) {
644 _json["level"] = level; 690 _json["level"] = level;
645 } 691 }
646 if (segment != null) { 692 if (segment != null) {
647 _json["segment"] = (segment).toJson(); 693 _json["segment"] = (segment).toJson();
648 } 694 }
649 return _json; 695 return _json;
650 } 696 }
651 } 697 }
652 698
653 /** 699 /// Safe search annotation (based on per-frame visual signals only).
654 * Safe search annotation (based on per-frame visual signals only). 700 /// If no unsafe content has been detected in a frame, no annotations
655 * If no unsafe content has been detected in a frame, no annotations 701 /// are present for that frame. If only some types of unsafe content
656 * are present for that frame. If only some types of unsafe content 702 /// have been detected in a frame, the likelihood is set to `UNKNOWN`
657 * have been detected in a frame, the likelihood is set to `UNKNOWN` 703 /// for all other types of unsafe content.
658 * for all other types of unsafe content.
659 */
660 class GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation { 704 class GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation {
661 /** 705 /// Likelihood of adult content.
662 * Likelihood of adult content. 706 /// Possible string values are:
663 * Possible string values are: 707 /// - "UNKNOWN" : Unknown likelihood.
664 * - "UNKNOWN" : Unknown likelihood. 708 /// - "VERY_UNLIKELY" : Very unlikely.
665 * - "VERY_UNLIKELY" : Very unlikely. 709 /// - "UNLIKELY" : Unlikely.
666 * - "UNLIKELY" : Unlikely. 710 /// - "POSSIBLE" : Possible.
667 * - "POSSIBLE" : Possible. 711 /// - "LIKELY" : Likely.
668 * - "LIKELY" : Likely. 712 /// - "VERY_LIKELY" : Very likely.
669 * - "VERY_LIKELY" : Very likely.
670 */
671 core.String adult; 713 core.String adult;
672 /** 714
673 * Likelihood of medical content. 715 /// Likelihood of medical content.
674 * Possible string values are: 716 /// Possible string values are:
675 * - "UNKNOWN" : Unknown likelihood. 717 /// - "UNKNOWN" : Unknown likelihood.
676 * - "VERY_UNLIKELY" : Very unlikely. 718 /// - "VERY_UNLIKELY" : Very unlikely.
677 * - "UNLIKELY" : Unlikely. 719 /// - "UNLIKELY" : Unlikely.
678 * - "POSSIBLE" : Possible. 720 /// - "POSSIBLE" : Possible.
679 * - "LIKELY" : Likely. 721 /// - "LIKELY" : Likely.
680 * - "VERY_LIKELY" : Very likely. 722 /// - "VERY_LIKELY" : Very likely.
681 */
682 core.String medical; 723 core.String medical;
683 /** 724
684 * Likelihood of racy content. 725 /// Likelihood of racy content.
685 * Possible string values are: 726 /// Possible string values are:
686 * - "UNKNOWN" : Unknown likelihood. 727 /// - "UNKNOWN" : Unknown likelihood.
687 * - "VERY_UNLIKELY" : Very unlikely. 728 /// - "VERY_UNLIKELY" : Very unlikely.
688 * - "UNLIKELY" : Unlikely. 729 /// - "UNLIKELY" : Unlikely.
689 * - "POSSIBLE" : Possible. 730 /// - "POSSIBLE" : Possible.
690 * - "LIKELY" : Likely. 731 /// - "LIKELY" : Likely.
691 * - "VERY_LIKELY" : Very likely. 732 /// - "VERY_LIKELY" : Very likely.
692 */
693 core.String racy; 733 core.String racy;
694 /** 734
695 * Likelihood that an obvious modification was made to the original 735 /// Likelihood that an obvious modification was made to the original
696 * version to make it appear funny or offensive. 736 /// version to make it appear funny or offensive.
697 * Possible string values are: 737 /// Possible string values are:
698 * - "UNKNOWN" : Unknown likelihood. 738 /// - "UNKNOWN" : Unknown likelihood.
699 * - "VERY_UNLIKELY" : Very unlikely. 739 /// - "VERY_UNLIKELY" : Very unlikely.
700 * - "UNLIKELY" : Unlikely. 740 /// - "UNLIKELY" : Unlikely.
701 * - "POSSIBLE" : Possible. 741 /// - "POSSIBLE" : Possible.
702 * - "LIKELY" : Likely. 742 /// - "LIKELY" : Likely.
703 * - "VERY_LIKELY" : Very likely. 743 /// - "VERY_LIKELY" : Very likely.
704 */
705 core.String spoof; 744 core.String spoof;
706 /** Video time offset in microseconds. */ 745
746 /// Video time offset in microseconds.
707 core.String timeOffset; 747 core.String timeOffset;
708 /** 748
709 * Likelihood of violent content. 749 /// Likelihood of violent content.
710 * Possible string values are: 750 /// Possible string values are:
711 * - "UNKNOWN" : Unknown likelihood. 751 /// - "UNKNOWN" : Unknown likelihood.
712 * - "VERY_UNLIKELY" : Very unlikely. 752 /// - "VERY_UNLIKELY" : Very unlikely.
713 * - "UNLIKELY" : Unlikely. 753 /// - "UNLIKELY" : Unlikely.
714 * - "POSSIBLE" : Possible. 754 /// - "POSSIBLE" : Possible.
715 * - "LIKELY" : Likely. 755 /// - "LIKELY" : Likely.
716 * - "VERY_LIKELY" : Very likely. 756 /// - "VERY_LIKELY" : Very likely.
717 */
718 core.String violent; 757 core.String violent;
719 758
720 GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation(); 759 GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation();
721 760
722 GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation.fromJson(core.Map _jso n) { 761 GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation.fromJson(
762 core.Map _json) {
723 if (_json.containsKey("adult")) { 763 if (_json.containsKey("adult")) {
724 adult = _json["adult"]; 764 adult = _json["adult"];
725 } 765 }
726 if (_json.containsKey("medical")) { 766 if (_json.containsKey("medical")) {
727 medical = _json["medical"]; 767 medical = _json["medical"];
728 } 768 }
729 if (_json.containsKey("racy")) { 769 if (_json.containsKey("racy")) {
730 racy = _json["racy"]; 770 racy = _json["racy"];
731 } 771 }
732 if (_json.containsKey("spoof")) { 772 if (_json.containsKey("spoof")) {
733 spoof = _json["spoof"]; 773 spoof = _json["spoof"];
734 } 774 }
735 if (_json.containsKey("timeOffset")) { 775 if (_json.containsKey("timeOffset")) {
736 timeOffset = _json["timeOffset"]; 776 timeOffset = _json["timeOffset"];
737 } 777 }
738 if (_json.containsKey("violent")) { 778 if (_json.containsKey("violent")) {
739 violent = _json["violent"]; 779 violent = _json["violent"];
740 } 780 }
741 } 781 }
742 782
743 core.Map<core.String, core.Object> toJson() { 783 core.Map<core.String, core.Object> toJson() {
744 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 784 final core.Map<core.String, core.Object> _json =
785 new core.Map<core.String, core.Object>();
745 if (adult != null) { 786 if (adult != null) {
746 _json["adult"] = adult; 787 _json["adult"] = adult;
747 } 788 }
748 if (medical != null) { 789 if (medical != null) {
749 _json["medical"] = medical; 790 _json["medical"] = medical;
750 } 791 }
751 if (racy != null) { 792 if (racy != null) {
752 _json["racy"] = racy; 793 _json["racy"] = racy;
753 } 794 }
754 if (spoof != null) { 795 if (spoof != null) {
755 _json["spoof"] = spoof; 796 _json["spoof"] = spoof;
756 } 797 }
757 if (timeOffset != null) { 798 if (timeOffset != null) {
758 _json["timeOffset"] = timeOffset; 799 _json["timeOffset"] = timeOffset;
759 } 800 }
760 if (violent != null) { 801 if (violent != null) {
761 _json["violent"] = violent; 802 _json["violent"] = violent;
762 } 803 }
763 return _json; 804 return _json;
764 } 805 }
765 } 806 }
766 807
767 /** Annotation progress for a single video. */ 808 /// Annotation progress for a single video.
768 class GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress { 809 class GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress {
769 /** 810 /// Video file location in
770 * Video file location in 811 /// [Google Cloud Storage](https://cloud.google.com/storage/).
771 * [Google Cloud Storage](https://cloud.google.com/storage/).
772 */
773 core.String inputUri; 812 core.String inputUri;
774 /** 813
775 * Approximate percentage processed thus far. 814 /// Approximate percentage processed thus far.
776 * Guaranteed to be 100 when fully processed. 815 /// Guaranteed to be 100 when fully processed.
777 */
778 core.int progressPercent; 816 core.int progressPercent;
779 /** Time when the request was received. */ 817
818 /// Time when the request was received.
780 core.String startTime; 819 core.String startTime;
781 /** Time of the most recent update. */ 820
821 /// Time of the most recent update.
782 core.String updateTime; 822 core.String updateTime;
783 823
784 GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress(); 824 GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress();
785 825
786 GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress.fromJson(core.Map _ json) { 826 GoogleCloudVideointelligenceV1beta1VideoAnnotationProgress.fromJson(
827 core.Map _json) {
787 if (_json.containsKey("inputUri")) { 828 if (_json.containsKey("inputUri")) {
788 inputUri = _json["inputUri"]; 829 inputUri = _json["inputUri"];
789 } 830 }
790 if (_json.containsKey("progressPercent")) { 831 if (_json.containsKey("progressPercent")) {
791 progressPercent = _json["progressPercent"]; 832 progressPercent = _json["progressPercent"];
792 } 833 }
793 if (_json.containsKey("startTime")) { 834 if (_json.containsKey("startTime")) {
794 startTime = _json["startTime"]; 835 startTime = _json["startTime"];
795 } 836 }
796 if (_json.containsKey("updateTime")) { 837 if (_json.containsKey("updateTime")) {
797 updateTime = _json["updateTime"]; 838 updateTime = _json["updateTime"];
798 } 839 }
799 } 840 }
800 841
801 core.Map<core.String, core.Object> toJson() { 842 core.Map<core.String, core.Object> toJson() {
802 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 843 final core.Map<core.String, core.Object> _json =
844 new core.Map<core.String, core.Object>();
803 if (inputUri != null) { 845 if (inputUri != null) {
804 _json["inputUri"] = inputUri; 846 _json["inputUri"] = inputUri;
805 } 847 }
806 if (progressPercent != null) { 848 if (progressPercent != null) {
807 _json["progressPercent"] = progressPercent; 849 _json["progressPercent"] = progressPercent;
808 } 850 }
809 if (startTime != null) { 851 if (startTime != null) {
810 _json["startTime"] = startTime; 852 _json["startTime"] = startTime;
811 } 853 }
812 if (updateTime != null) { 854 if (updateTime != null) {
813 _json["updateTime"] = updateTime; 855 _json["updateTime"] = updateTime;
814 } 856 }
815 return _json; 857 return _json;
816 } 858 }
817 } 859 }
818 860
819 /** Annotation results for a single video. */ 861 /// Annotation results for a single video.
820 class GoogleCloudVideointelligenceV1beta1VideoAnnotationResults { 862 class GoogleCloudVideointelligenceV1beta1VideoAnnotationResults {
821 /** 863 /// If set, indicates an error. Note that for a single `AnnotateVideoRequest`
822 * If set, indicates an error. Note that for a single `AnnotateVideoRequest` 864 /// some videos may succeed and some may fail.
823 * some videos may succeed and some may fail.
824 */
825 GoogleRpcStatus error; 865 GoogleRpcStatus error;
826 /** 866
827 * Video file location in 867 /// Video file location in
828 * [Google Cloud Storage](https://cloud.google.com/storage/). 868 /// [Google Cloud Storage](https://cloud.google.com/storage/).
829 */
830 core.String inputUri; 869 core.String inputUri;
831 /** Label annotations. There is exactly one element for each unique label. */ 870
832 core.List<GoogleCloudVideointelligenceV1beta1LabelAnnotation> labelAnnotations ; 871 /// Label annotations. There is exactly one element for each unique label.
833 /** Safe search annotations. */ 872 core.List<GoogleCloudVideointelligenceV1beta1LabelAnnotation>
834 core.List<GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation> safeSearchA nnotations; 873 labelAnnotations;
835 /** Shot annotations. Each shot is represented as a video segment. */ 874
875 /// Safe search annotations.
876 core.List<GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation>
877 safeSearchAnnotations;
878
879 /// Shot annotations. Each shot is represented as a video segment.
836 core.List<GoogleCloudVideointelligenceV1beta1VideoSegment> shotAnnotations; 880 core.List<GoogleCloudVideointelligenceV1beta1VideoSegment> shotAnnotations;
837 881
838 GoogleCloudVideointelligenceV1beta1VideoAnnotationResults(); 882 GoogleCloudVideointelligenceV1beta1VideoAnnotationResults();
839 883
840 GoogleCloudVideointelligenceV1beta1VideoAnnotationResults.fromJson(core.Map _j son) { 884 GoogleCloudVideointelligenceV1beta1VideoAnnotationResults.fromJson(
885 core.Map _json) {
841 if (_json.containsKey("error")) { 886 if (_json.containsKey("error")) {
842 error = new GoogleRpcStatus.fromJson(_json["error"]); 887 error = new GoogleRpcStatus.fromJson(_json["error"]);
843 } 888 }
844 if (_json.containsKey("inputUri")) { 889 if (_json.containsKey("inputUri")) {
845 inputUri = _json["inputUri"]; 890 inputUri = _json["inputUri"];
846 } 891 }
847 if (_json.containsKey("labelAnnotations")) { 892 if (_json.containsKey("labelAnnotations")) {
848 labelAnnotations = _json["labelAnnotations"].map((value) => new GoogleClou dVideointelligenceV1beta1LabelAnnotation.fromJson(value)).toList(); 893 labelAnnotations = _json["labelAnnotations"]
894 .map((value) =>
895 new GoogleCloudVideointelligenceV1beta1LabelAnnotation.fromJson(
896 value))
897 .toList();
849 } 898 }
850 if (_json.containsKey("safeSearchAnnotations")) { 899 if (_json.containsKey("safeSearchAnnotations")) {
851 safeSearchAnnotations = _json["safeSearchAnnotations"].map((value) => new GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation.fromJson(value)).toList( ); 900 safeSearchAnnotations = _json["safeSearchAnnotations"]
901 .map((value) =>
902 new GoogleCloudVideointelligenceV1beta1SafeSearchAnnotation
903 .fromJson(value))
904 .toList();
852 } 905 }
853 if (_json.containsKey("shotAnnotations")) { 906 if (_json.containsKey("shotAnnotations")) {
854 shotAnnotations = _json["shotAnnotations"].map((value) => new GoogleCloudV ideointelligenceV1beta1VideoSegment.fromJson(value)).toList(); 907 shotAnnotations = _json["shotAnnotations"]
908 .map((value) =>
909 new GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(
910 value))
911 .toList();
855 } 912 }
856 } 913 }
857 914
858 core.Map<core.String, core.Object> toJson() { 915 core.Map<core.String, core.Object> toJson() {
859 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 916 final core.Map<core.String, core.Object> _json =
917 new core.Map<core.String, core.Object>();
860 if (error != null) { 918 if (error != null) {
861 _json["error"] = (error).toJson(); 919 _json["error"] = (error).toJson();
862 } 920 }
863 if (inputUri != null) { 921 if (inputUri != null) {
864 _json["inputUri"] = inputUri; 922 _json["inputUri"] = inputUri;
865 } 923 }
866 if (labelAnnotations != null) { 924 if (labelAnnotations != null) {
867 _json["labelAnnotations"] = labelAnnotations.map((value) => (value).toJson ()).toList(); 925 _json["labelAnnotations"] =
926 labelAnnotations.map((value) => (value).toJson()).toList();
868 } 927 }
869 if (safeSearchAnnotations != null) { 928 if (safeSearchAnnotations != null) {
870 _json["safeSearchAnnotations"] = safeSearchAnnotations.map((value) => (val ue).toJson()).toList(); 929 _json["safeSearchAnnotations"] =
930 safeSearchAnnotations.map((value) => (value).toJson()).toList();
871 } 931 }
872 if (shotAnnotations != null) { 932 if (shotAnnotations != null) {
873 _json["shotAnnotations"] = shotAnnotations.map((value) => (value).toJson() ).toList(); 933 _json["shotAnnotations"] =
934 shotAnnotations.map((value) => (value).toJson()).toList();
874 } 935 }
875 return _json; 936 return _json;
876 } 937 }
877 } 938 }
878 939
879 /** Video context and/or feature-specific parameters. */ 940 /// Video context and/or feature-specific parameters.
880 class GoogleCloudVideointelligenceV1beta1VideoContext { 941 class GoogleCloudVideointelligenceV1beta1VideoContext {
881 /** 942 /// If label detection has been requested, what labels should be detected
882 * If label detection has been requested, what labels should be detected 943 /// in addition to video-level labels or segment-level labels. If
883 * in addition to video-level labels or segment-level labels. If unspecified, 944 /// unspecified,
884 * defaults to `SHOT_MODE`. 945 /// defaults to `SHOT_MODE`.
885 * Possible string values are: 946 /// Possible string values are:
886 * - "LABEL_DETECTION_MODE_UNSPECIFIED" : Unspecified. 947 /// - "LABEL_DETECTION_MODE_UNSPECIFIED" : Unspecified.
887 * - "SHOT_MODE" : Detect shot-level labels. 948 /// - "SHOT_MODE" : Detect shot-level labels.
888 * - "FRAME_MODE" : Detect frame-level labels. 949 /// - "FRAME_MODE" : Detect frame-level labels.
889 * - "SHOT_AND_FRAME_MODE" : Detect both shot-level and frame-level labels. 950 /// - "SHOT_AND_FRAME_MODE" : Detect both shot-level and frame-level labels.
890 */
891 core.String labelDetectionMode; 951 core.String labelDetectionMode;
892 /** 952
893 * Model to use for label detection. 953 /// Model to use for label detection.
894 * Supported values: "latest" and "stable" (the default). 954 /// Supported values: "latest" and "stable" (the default).
895 */
896 core.String labelDetectionModel; 955 core.String labelDetectionModel;
897 /** 956
898 * Model to use for safe search detection. 957 /// Model to use for safe search detection.
899 * Supported values: "latest" and "stable" (the default). 958 /// Supported values: "latest" and "stable" (the default).
900 */
901 core.String safeSearchDetectionModel; 959 core.String safeSearchDetectionModel;
902 /** 960
903 * Video segments to annotate. The segments may overlap and are not required 961 /// Video segments to annotate. The segments may overlap and are not required
904 * to be contiguous or span the whole video. If unspecified, each video 962 /// to be contiguous or span the whole video. If unspecified, each video
905 * is treated as a single segment. 963 /// is treated as a single segment.
906 */
907 core.List<GoogleCloudVideointelligenceV1beta1VideoSegment> segments; 964 core.List<GoogleCloudVideointelligenceV1beta1VideoSegment> segments;
908 /** 965
909 * Model to use for shot change detection. 966 /// Model to use for shot change detection.
910 * Supported values: "latest" and "stable" (the default). 967 /// Supported values: "latest" and "stable" (the default).
911 */
912 core.String shotChangeDetectionModel; 968 core.String shotChangeDetectionModel;
913 /** 969
914 * Whether the video has been shot from a stationary (i.e. non-moving) camera. 970 /// Whether the video has been shot from a stationary (i.e. non-moving)
915 * When set to true, might improve detection accuracy for moving objects. 971 /// camera.
916 */ 972 /// When set to true, might improve detection accuracy for moving objects.
917 core.bool stationaryCamera; 973 core.bool stationaryCamera;
918 974
919 GoogleCloudVideointelligenceV1beta1VideoContext(); 975 GoogleCloudVideointelligenceV1beta1VideoContext();
920 976
921 GoogleCloudVideointelligenceV1beta1VideoContext.fromJson(core.Map _json) { 977 GoogleCloudVideointelligenceV1beta1VideoContext.fromJson(core.Map _json) {
922 if (_json.containsKey("labelDetectionMode")) { 978 if (_json.containsKey("labelDetectionMode")) {
923 labelDetectionMode = _json["labelDetectionMode"]; 979 labelDetectionMode = _json["labelDetectionMode"];
924 } 980 }
925 if (_json.containsKey("labelDetectionModel")) { 981 if (_json.containsKey("labelDetectionModel")) {
926 labelDetectionModel = _json["labelDetectionModel"]; 982 labelDetectionModel = _json["labelDetectionModel"];
927 } 983 }
928 if (_json.containsKey("safeSearchDetectionModel")) { 984 if (_json.containsKey("safeSearchDetectionModel")) {
929 safeSearchDetectionModel = _json["safeSearchDetectionModel"]; 985 safeSearchDetectionModel = _json["safeSearchDetectionModel"];
930 } 986 }
931 if (_json.containsKey("segments")) { 987 if (_json.containsKey("segments")) {
932 segments = _json["segments"].map((value) => new GoogleCloudVideointelligen ceV1beta1VideoSegment.fromJson(value)).toList(); 988 segments = _json["segments"]
989 .map((value) =>
990 new GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(
991 value))
992 .toList();
933 } 993 }
934 if (_json.containsKey("shotChangeDetectionModel")) { 994 if (_json.containsKey("shotChangeDetectionModel")) {
935 shotChangeDetectionModel = _json["shotChangeDetectionModel"]; 995 shotChangeDetectionModel = _json["shotChangeDetectionModel"];
936 } 996 }
937 if (_json.containsKey("stationaryCamera")) { 997 if (_json.containsKey("stationaryCamera")) {
938 stationaryCamera = _json["stationaryCamera"]; 998 stationaryCamera = _json["stationaryCamera"];
939 } 999 }
940 } 1000 }
941 1001
942 core.Map<core.String, core.Object> toJson() { 1002 core.Map<core.String, core.Object> toJson() {
943 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 1003 final core.Map<core.String, core.Object> _json =
1004 new core.Map<core.String, core.Object>();
944 if (labelDetectionMode != null) { 1005 if (labelDetectionMode != null) {
945 _json["labelDetectionMode"] = labelDetectionMode; 1006 _json["labelDetectionMode"] = labelDetectionMode;
946 } 1007 }
947 if (labelDetectionModel != null) { 1008 if (labelDetectionModel != null) {
948 _json["labelDetectionModel"] = labelDetectionModel; 1009 _json["labelDetectionModel"] = labelDetectionModel;
949 } 1010 }
950 if (safeSearchDetectionModel != null) { 1011 if (safeSearchDetectionModel != null) {
951 _json["safeSearchDetectionModel"] = safeSearchDetectionModel; 1012 _json["safeSearchDetectionModel"] = safeSearchDetectionModel;
952 } 1013 }
953 if (segments != null) { 1014 if (segments != null) {
954 _json["segments"] = segments.map((value) => (value).toJson()).toList(); 1015 _json["segments"] = segments.map((value) => (value).toJson()).toList();
955 } 1016 }
956 if (shotChangeDetectionModel != null) { 1017 if (shotChangeDetectionModel != null) {
957 _json["shotChangeDetectionModel"] = shotChangeDetectionModel; 1018 _json["shotChangeDetectionModel"] = shotChangeDetectionModel;
958 } 1019 }
959 if (stationaryCamera != null) { 1020 if (stationaryCamera != null) {
960 _json["stationaryCamera"] = stationaryCamera; 1021 _json["stationaryCamera"] = stationaryCamera;
961 } 1022 }
962 return _json; 1023 return _json;
963 } 1024 }
964 } 1025 }
965 1026
966 /** Video segment. */ 1027 /// Video segment.
967 class GoogleCloudVideointelligenceV1beta1VideoSegment { 1028 class GoogleCloudVideointelligenceV1beta1VideoSegment {
968 /** End offset in microseconds (inclusive). Unset means 0. */ 1029 /// End offset in microseconds (inclusive). Unset means 0.
969 core.String endTimeOffset; 1030 core.String endTimeOffset;
970 /** Start offset in microseconds (inclusive). Unset means 0. */ 1031
1032 /// Start offset in microseconds (inclusive). Unset means 0.
971 core.String startTimeOffset; 1033 core.String startTimeOffset;
972 1034
973 GoogleCloudVideointelligenceV1beta1VideoSegment(); 1035 GoogleCloudVideointelligenceV1beta1VideoSegment();
974 1036
975 GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(core.Map _json) { 1037 GoogleCloudVideointelligenceV1beta1VideoSegment.fromJson(core.Map _json) {
976 if (_json.containsKey("endTimeOffset")) { 1038 if (_json.containsKey("endTimeOffset")) {
977 endTimeOffset = _json["endTimeOffset"]; 1039 endTimeOffset = _json["endTimeOffset"];
978 } 1040 }
979 if (_json.containsKey("startTimeOffset")) { 1041 if (_json.containsKey("startTimeOffset")) {
980 startTimeOffset = _json["startTimeOffset"]; 1042 startTimeOffset = _json["startTimeOffset"];
981 } 1043 }
982 } 1044 }
983 1045
984 core.Map<core.String, core.Object> toJson() { 1046 core.Map<core.String, core.Object> toJson() {
985 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 1047 final core.Map<core.String, core.Object> _json =
1048 new core.Map<core.String, core.Object>();
986 if (endTimeOffset != null) { 1049 if (endTimeOffset != null) {
987 _json["endTimeOffset"] = endTimeOffset; 1050 _json["endTimeOffset"] = endTimeOffset;
988 } 1051 }
1052 if (startTimeOffset != null) {
1053 _json["startTimeOffset"] = startTimeOffset;
1054 }
1055 return _json;
1056 }
1057 }
1058
1059 /// Video annotation progress. Included in the `metadata`
1060 /// field of the `Operation` returned by the `GetOperation`
1061 /// call of the `google::longrunning::Operations` service.
1062 class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress {
1063 /// Progress metadata for all videos specified in `AnnotateVideoRequest`.
1064 core.List<GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress>
1065 annotationProgress;
1066
1067 GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress();
1068
1069 GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress.fromJson(
1070 core.Map _json) {
1071 if (_json.containsKey("annotationProgress")) {
1072 annotationProgress = _json["annotationProgress"]
1073 .map((value) =>
1074 new GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress
1075 .fromJson(value))
1076 .toList();
1077 }
1078 }
1079
1080 core.Map<core.String, core.Object> toJson() {
1081 final core.Map<core.String, core.Object> _json =
1082 new core.Map<core.String, core.Object>();
1083 if (annotationProgress != null) {
1084 _json["annotationProgress"] =
1085 annotationProgress.map((value) => (value).toJson()).toList();
1086 }
1087 return _json;
1088 }
1089 }
1090
1091 /// Video annotation response. Included in the `response`
1092 /// field of the `Operation` returned by the `GetOperation`
1093 /// call of the `google::longrunning::Operations` service.
1094 class GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse {
1095 /// Annotation results for all videos specified in `AnnotateVideoRequest`.
1096 core.List<GoogleCloudVideointelligenceV1beta2VideoAnnotationResults>
1097 annotationResults;
1098
1099 GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse();
1100
1101 GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse.fromJson(
1102 core.Map _json) {
1103 if (_json.containsKey("annotationResults")) {
1104 annotationResults = _json["annotationResults"]
1105 .map((value) =>
1106 new GoogleCloudVideointelligenceV1beta2VideoAnnotationResults
1107 .fromJson(value))
1108 .toList();
1109 }
1110 }
1111
1112 core.Map<core.String, core.Object> toJson() {
1113 final core.Map<core.String, core.Object> _json =
1114 new core.Map<core.String, core.Object>();
1115 if (annotationResults != null) {
1116 _json["annotationResults"] =
1117 annotationResults.map((value) => (value).toJson()).toList();
1118 }
1119 return _json;
1120 }
1121 }
1122
1123 /// Detected entity from video analysis.
1124 class GoogleCloudVideointelligenceV1beta2Entity {
1125 /// Textual description, e.g. `Fixed-gear bicycle`.
1126 core.String description;
1127
1128 /// Opaque entity ID. Some IDs may be available in
1129 /// [Google Knowledge Graph Search
1130 /// API](https://developers.google.com/knowledge-graph/).
1131 core.String entityId;
1132
1133 /// Language code for `description` in BCP-47 format.
1134 core.String languageCode;
1135
1136 GoogleCloudVideointelligenceV1beta2Entity();
1137
1138 GoogleCloudVideointelligenceV1beta2Entity.fromJson(core.Map _json) {
1139 if (_json.containsKey("description")) {
1140 description = _json["description"];
1141 }
1142 if (_json.containsKey("entityId")) {
1143 entityId = _json["entityId"];
1144 }
1145 if (_json.containsKey("languageCode")) {
1146 languageCode = _json["languageCode"];
1147 }
1148 }
1149
1150 core.Map<core.String, core.Object> toJson() {
1151 final core.Map<core.String, core.Object> _json =
1152 new core.Map<core.String, core.Object>();
1153 if (description != null) {
1154 _json["description"] = description;
1155 }
1156 if (entityId != null) {
1157 _json["entityId"] = entityId;
1158 }
1159 if (languageCode != null) {
1160 _json["languageCode"] = languageCode;
1161 }
1162 return _json;
1163 }
1164 }
1165
1166 /// Explicit content annotation (based on per-frame visual signals only).
1167 /// If no explicit content has been detected in a frame, no annotations are
1168 /// present for that frame.
1169 class GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation {
1170 /// All video frames where explicit content was detected.
1171 core.List<GoogleCloudVideointelligenceV1beta2ExplicitContentFrame> frames;
1172
1173 GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation();
1174
1175 GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation.fromJson(
1176 core.Map _json) {
1177 if (_json.containsKey("frames")) {
1178 frames = _json["frames"]
1179 .map((value) =>
1180 new GoogleCloudVideointelligenceV1beta2ExplicitContentFrame
1181 .fromJson(value))
1182 .toList();
1183 }
1184 }
1185
1186 core.Map<core.String, core.Object> toJson() {
1187 final core.Map<core.String, core.Object> _json =
1188 new core.Map<core.String, core.Object>();
1189 if (frames != null) {
1190 _json["frames"] = frames.map((value) => (value).toJson()).toList();
1191 }
1192 return _json;
1193 }
1194 }
1195
1196 /// Video frame level annotation results for explicit content.
1197 class GoogleCloudVideointelligenceV1beta2ExplicitContentFrame {
1198 /// Likelihood of the pornography content..
1199 /// Possible string values are:
1200 /// - "LIKELIHOOD_UNSPECIFIED" : Unspecified likelihood.
1201 /// - "VERY_UNLIKELY" : Very unlikely.
1202 /// - "UNLIKELY" : Unlikely.
1203 /// - "POSSIBLE" : Possible.
1204 /// - "LIKELY" : Likely.
1205 /// - "VERY_LIKELY" : Very likely.
1206 core.String pornographyLikelihood;
1207
1208 /// Time-offset, relative to the beginning of the video, corresponding to the
1209 /// video frame for this location.
1210 core.String timeOffset;
1211
1212 GoogleCloudVideointelligenceV1beta2ExplicitContentFrame();
1213
1214 GoogleCloudVideointelligenceV1beta2ExplicitContentFrame.fromJson(
1215 core.Map _json) {
1216 if (_json.containsKey("pornographyLikelihood")) {
1217 pornographyLikelihood = _json["pornographyLikelihood"];
1218 }
1219 if (_json.containsKey("timeOffset")) {
1220 timeOffset = _json["timeOffset"];
1221 }
1222 }
1223
1224 core.Map<core.String, core.Object> toJson() {
1225 final core.Map<core.String, core.Object> _json =
1226 new core.Map<core.String, core.Object>();
1227 if (pornographyLikelihood != null) {
1228 _json["pornographyLikelihood"] = pornographyLikelihood;
1229 }
1230 if (timeOffset != null) {
1231 _json["timeOffset"] = timeOffset;
1232 }
1233 return _json;
1234 }
1235 }
1236
1237 /// Label annotation.
1238 class GoogleCloudVideointelligenceV1beta2LabelAnnotation {
1239 /// Common categories for the detected entity.
1240 /// E.g. when the label is `Terrier` the category is likely `dog`. And in
1241 /// some
1242 /// cases there might be more than one categories e.g. `Terrier` could also
1243 /// be
1244 /// a `pet`.
1245 core.List<GoogleCloudVideointelligenceV1beta2Entity> categoryEntities;
1246
1247 /// Detected entity.
1248 GoogleCloudVideointelligenceV1beta2Entity entity;
1249
1250 /// All video frames where a label was detected.
1251 core.List<GoogleCloudVideointelligenceV1beta2LabelFrame> frames;
1252
1253 /// All video segments where a label was detected.
1254 core.List<GoogleCloudVideointelligenceV1beta2LabelSegment> segments;
1255
1256 GoogleCloudVideointelligenceV1beta2LabelAnnotation();
1257
1258 GoogleCloudVideointelligenceV1beta2LabelAnnotation.fromJson(core.Map _json) {
1259 if (_json.containsKey("categoryEntities")) {
1260 categoryEntities = _json["categoryEntities"]
1261 .map((value) =>
1262 new GoogleCloudVideointelligenceV1beta2Entity.fromJson(value))
1263 .toList();
1264 }
1265 if (_json.containsKey("entity")) {
1266 entity = new GoogleCloudVideointelligenceV1beta2Entity.fromJson(
1267 _json["entity"]);
1268 }
1269 if (_json.containsKey("frames")) {
1270 frames = _json["frames"]
1271 .map((value) =>
1272 new GoogleCloudVideointelligenceV1beta2LabelFrame.fromJson(value))
1273 .toList();
1274 }
1275 if (_json.containsKey("segments")) {
1276 segments = _json["segments"]
1277 .map((value) =>
1278 new GoogleCloudVideointelligenceV1beta2LabelSegment.fromJson(
1279 value))
1280 .toList();
1281 }
1282 }
1283
1284 core.Map<core.String, core.Object> toJson() {
1285 final core.Map<core.String, core.Object> _json =
1286 new core.Map<core.String, core.Object>();
1287 if (categoryEntities != null) {
1288 _json["categoryEntities"] =
1289 categoryEntities.map((value) => (value).toJson()).toList();
1290 }
1291 if (entity != null) {
1292 _json["entity"] = (entity).toJson();
1293 }
1294 if (frames != null) {
1295 _json["frames"] = frames.map((value) => (value).toJson()).toList();
1296 }
1297 if (segments != null) {
1298 _json["segments"] = segments.map((value) => (value).toJson()).toList();
1299 }
1300 return _json;
1301 }
1302 }
1303
1304 /// Video frame level annotation results for label detection.
1305 class GoogleCloudVideointelligenceV1beta2LabelFrame {
1306 /// Confidence that the label is accurate. Range: [0, 1].
1307 core.double confidence;
1308
1309 /// Time-offset, relative to the beginning of the video, corresponding to the
1310 /// video frame for this location.
1311 core.String timeOffset;
1312
1313 GoogleCloudVideointelligenceV1beta2LabelFrame();
1314
1315 GoogleCloudVideointelligenceV1beta2LabelFrame.fromJson(core.Map _json) {
1316 if (_json.containsKey("confidence")) {
1317 confidence = _json["confidence"];
1318 }
1319 if (_json.containsKey("timeOffset")) {
1320 timeOffset = _json["timeOffset"];
1321 }
1322 }
1323
1324 core.Map<core.String, core.Object> toJson() {
1325 final core.Map<core.String, core.Object> _json =
1326 new core.Map<core.String, core.Object>();
1327 if (confidence != null) {
1328 _json["confidence"] = confidence;
1329 }
1330 if (timeOffset != null) {
1331 _json["timeOffset"] = timeOffset;
1332 }
1333 return _json;
1334 }
1335 }
1336
1337 /// Video segment level annotation results for label detection.
1338 class GoogleCloudVideointelligenceV1beta2LabelSegment {
1339 /// Confidence that the label is accurate. Range: [0, 1].
1340 core.double confidence;
1341
1342 /// Video segment where a label was detected.
1343 GoogleCloudVideointelligenceV1beta2VideoSegment segment;
1344
1345 GoogleCloudVideointelligenceV1beta2LabelSegment();
1346
1347 GoogleCloudVideointelligenceV1beta2LabelSegment.fromJson(core.Map _json) {
1348 if (_json.containsKey("confidence")) {
1349 confidence = _json["confidence"];
1350 }
1351 if (_json.containsKey("segment")) {
1352 segment = new GoogleCloudVideointelligenceV1beta2VideoSegment.fromJson(
1353 _json["segment"]);
1354 }
1355 }
1356
1357 core.Map<core.String, core.Object> toJson() {
1358 final core.Map<core.String, core.Object> _json =
1359 new core.Map<core.String, core.Object>();
1360 if (confidence != null) {
1361 _json["confidence"] = confidence;
1362 }
1363 if (segment != null) {
1364 _json["segment"] = (segment).toJson();
1365 }
1366 return _json;
1367 }
1368 }
1369
1370 /// Annotation progress for a single video.
1371 class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress {
1372 /// Video file location in
1373 /// [Google Cloud Storage](https://cloud.google.com/storage/).
1374 core.String inputUri;
1375
1376 /// Approximate percentage processed thus far.
1377 /// Guaranteed to be 100 when fully processed.
1378 core.int progressPercent;
1379
1380 /// Time when the request was received.
1381 core.String startTime;
1382
1383 /// Time of the most recent update.
1384 core.String updateTime;
1385
1386 GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress();
1387
1388 GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress.fromJson(
1389 core.Map _json) {
1390 if (_json.containsKey("inputUri")) {
1391 inputUri = _json["inputUri"];
1392 }
1393 if (_json.containsKey("progressPercent")) {
1394 progressPercent = _json["progressPercent"];
1395 }
1396 if (_json.containsKey("startTime")) {
1397 startTime = _json["startTime"];
1398 }
1399 if (_json.containsKey("updateTime")) {
1400 updateTime = _json["updateTime"];
1401 }
1402 }
1403
1404 core.Map<core.String, core.Object> toJson() {
1405 final core.Map<core.String, core.Object> _json =
1406 new core.Map<core.String, core.Object>();
1407 if (inputUri != null) {
1408 _json["inputUri"] = inputUri;
1409 }
1410 if (progressPercent != null) {
1411 _json["progressPercent"] = progressPercent;
1412 }
1413 if (startTime != null) {
1414 _json["startTime"] = startTime;
1415 }
1416 if (updateTime != null) {
1417 _json["updateTime"] = updateTime;
1418 }
1419 return _json;
1420 }
1421 }
1422
1423 /// Annotation results for a single video.
1424 class GoogleCloudVideointelligenceV1beta2VideoAnnotationResults {
1425 /// If set, indicates an error. Note that for a single `AnnotateVideoRequest`
1426 /// some videos may succeed and some may fail.
1427 GoogleRpcStatus error;
1428
1429 /// Explicit content annotation.
1430 GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation
1431 explicitAnnotation;
1432
1433 /// Label annotations on frame level.
1434 /// There is exactly one element for each unique label.
1435 core.List<GoogleCloudVideointelligenceV1beta2LabelAnnotation>
1436 frameLabelAnnotations;
1437
1438 /// Video file location in
1439 /// [Google Cloud Storage](https://cloud.google.com/storage/).
1440 core.String inputUri;
1441
1442 /// Label annotations on video level or user specified segment level.
1443 /// There is exactly one element for each unique label.
1444 core.List<GoogleCloudVideointelligenceV1beta2LabelAnnotation>
1445 segmentLabelAnnotations;
1446
1447 /// Shot annotations. Each shot is represented as a video segment.
1448 core.List<GoogleCloudVideointelligenceV1beta2VideoSegment> shotAnnotations;
1449
1450 /// Label annotations on shot level.
1451 /// There is exactly one element for each unique label.
1452 core.List<GoogleCloudVideointelligenceV1beta2LabelAnnotation>
1453 shotLabelAnnotations;
1454
1455 GoogleCloudVideointelligenceV1beta2VideoAnnotationResults();
1456
1457 GoogleCloudVideointelligenceV1beta2VideoAnnotationResults.fromJson(
1458 core.Map _json) {
1459 if (_json.containsKey("error")) {
1460 error = new GoogleRpcStatus.fromJson(_json["error"]);
1461 }
1462 if (_json.containsKey("explicitAnnotation")) {
1463 explicitAnnotation =
1464 new GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation
1465 .fromJson(_json["explicitAnnotation"]);
1466 }
1467 if (_json.containsKey("frameLabelAnnotations")) {
1468 frameLabelAnnotations = _json["frameLabelAnnotations"]
1469 .map((value) =>
1470 new GoogleCloudVideointelligenceV1beta2LabelAnnotation.fromJson(
1471 value))
1472 .toList();
1473 }
1474 if (_json.containsKey("inputUri")) {
1475 inputUri = _json["inputUri"];
1476 }
1477 if (_json.containsKey("segmentLabelAnnotations")) {
1478 segmentLabelAnnotations = _json["segmentLabelAnnotations"]
1479 .map((value) =>
1480 new GoogleCloudVideointelligenceV1beta2LabelAnnotation.fromJson(
1481 value))
1482 .toList();
1483 }
1484 if (_json.containsKey("shotAnnotations")) {
1485 shotAnnotations = _json["shotAnnotations"]
1486 .map((value) =>
1487 new GoogleCloudVideointelligenceV1beta2VideoSegment.fromJson(
1488 value))
1489 .toList();
1490 }
1491 if (_json.containsKey("shotLabelAnnotations")) {
1492 shotLabelAnnotations = _json["shotLabelAnnotations"]
1493 .map((value) =>
1494 new GoogleCloudVideointelligenceV1beta2LabelAnnotation.fromJson(
1495 value))
1496 .toList();
1497 }
1498 }
1499
1500 core.Map<core.String, core.Object> toJson() {
1501 final core.Map<core.String, core.Object> _json =
1502 new core.Map<core.String, core.Object>();
1503 if (error != null) {
1504 _json["error"] = (error).toJson();
1505 }
1506 if (explicitAnnotation != null) {
1507 _json["explicitAnnotation"] = (explicitAnnotation).toJson();
1508 }
1509 if (frameLabelAnnotations != null) {
1510 _json["frameLabelAnnotations"] =
1511 frameLabelAnnotations.map((value) => (value).toJson()).toList();
1512 }
1513 if (inputUri != null) {
1514 _json["inputUri"] = inputUri;
1515 }
1516 if (segmentLabelAnnotations != null) {
1517 _json["segmentLabelAnnotations"] =
1518 segmentLabelAnnotations.map((value) => (value).toJson()).toList();
1519 }
1520 if (shotAnnotations != null) {
1521 _json["shotAnnotations"] =
1522 shotAnnotations.map((value) => (value).toJson()).toList();
1523 }
1524 if (shotLabelAnnotations != null) {
1525 _json["shotLabelAnnotations"] =
1526 shotLabelAnnotations.map((value) => (value).toJson()).toList();
1527 }
1528 return _json;
1529 }
1530 }
1531
1532 /// Video segment.
1533 class GoogleCloudVideointelligenceV1beta2VideoSegment {
1534 /// Time-offset, relative to the beginning of the video,
1535 /// corresponding to the end of the segment (inclusive).
1536 core.String endTimeOffset;
1537
1538 /// Time-offset, relative to the beginning of the video,
1539 /// corresponding to the start of the segment (inclusive).
1540 core.String startTimeOffset;
1541
1542 GoogleCloudVideointelligenceV1beta2VideoSegment();
1543
1544 GoogleCloudVideointelligenceV1beta2VideoSegment.fromJson(core.Map _json) {
1545 if (_json.containsKey("endTimeOffset")) {
1546 endTimeOffset = _json["endTimeOffset"];
1547 }
1548 if (_json.containsKey("startTimeOffset")) {
1549 startTimeOffset = _json["startTimeOffset"];
1550 }
1551 }
1552
1553 core.Map<core.String, core.Object> toJson() {
1554 final core.Map<core.String, core.Object> _json =
1555 new core.Map<core.String, core.Object>();
1556 if (endTimeOffset != null) {
1557 _json["endTimeOffset"] = endTimeOffset;
1558 }
989 if (startTimeOffset != null) { 1559 if (startTimeOffset != null) {
990 _json["startTimeOffset"] = startTimeOffset; 1560 _json["startTimeOffset"] = startTimeOffset;
991 } 1561 }
992 return _json; 1562 return _json;
993 } 1563 }
994 } 1564 }
995 1565
996 /** 1566 /// This resource represents a long-running operation that is the result of a
997 * This resource represents a long-running operation that is the result of a 1567 /// network API call.
998 * network API call.
999 */
1000 class GoogleLongrunningOperation { 1568 class GoogleLongrunningOperation {
1001 /** 1569 /// If the value is `false`, it means the operation is still in progress.
1002 * If the value is `false`, it means the operation is still in progress. 1570 /// If `true`, the operation is completed, and either `error` or `response`
1003 * If true, the operation is completed, and either `error` or `response` is 1571 /// is
1004 * available. 1572 /// available.
1005 */
1006 core.bool done; 1573 core.bool done;
1007 /** The error result of the operation in case of failure or cancellation. */ 1574
1575 /// The error result of the operation in case of failure or cancellation.
1008 GoogleRpcStatus error; 1576 GoogleRpcStatus error;
1009 /** 1577
1010 * Service-specific metadata associated with the operation. It typically 1578 /// Service-specific metadata associated with the operation. It typically
1011 * contains progress information and common metadata such as create time. 1579 /// contains progress information and common metadata such as create time.
1012 * Some services might not provide such metadata. Any method that returns a 1580 /// Some services might not provide such metadata. Any method that returns a
1013 * long-running operation should document the metadata type, if any. 1581 /// long-running operation should document the metadata type, if any.
1014 * 1582 ///
1015 * The values for Object must be JSON objects. It can consist of `num`, 1583 /// The values for Object must be JSON objects. It can consist of `num`,
1016 * `String`, `bool` and `null` as well as `Map` and `List` values. 1584 /// `String`, `bool` and `null` as well as `Map` and `List` values.
1017 */
1018 core.Map<core.String, core.Object> metadata; 1585 core.Map<core.String, core.Object> metadata;
1019 /** 1586
1020 * The server-assigned name, which is only unique within the same service that 1587 /// The server-assigned name, which is only unique within the same service
1021 * originally returns it. If you use the default HTTP mapping, the 1588 /// that
1022 * `name` should have the format of `operations/some/unique/name`. 1589 /// originally returns it. If you use the default HTTP mapping, the
1023 */ 1590 /// `name` should have the format of `operations/some/unique/name`.
1024 core.String name; 1591 core.String name;
1025 /** 1592
1026 * The normal response of the operation in case of success. If the original 1593 /// The normal response of the operation in case of success. If the original
1027 * method returns no data on success, such as `Delete`, the response is 1594 /// method returns no data on success, such as `Delete`, the response is
1028 * `google.protobuf.Empty`. If the original method is standard 1595 /// `google.protobuf.Empty`. If the original method is standard
1029 * `Get`/`Create`/`Update`, the response should be the resource. For other 1596 /// `Get`/`Create`/`Update`, the response should be the resource. For other
1030 * methods, the response should have the type `XxxResponse`, where `Xxx` 1597 /// methods, the response should have the type `XxxResponse`, where `Xxx`
1031 * is the original method name. For example, if the original method name 1598 /// is the original method name. For example, if the original method name
1032 * is `TakeSnapshot()`, the inferred response type is 1599 /// is `TakeSnapshot()`, the inferred response type is
1033 * `TakeSnapshotResponse`. 1600 /// `TakeSnapshotResponse`.
1034 * 1601 ///
1035 * The values for Object must be JSON objects. It can consist of `num`, 1602 /// The values for Object must be JSON objects. It can consist of `num`,
1036 * `String`, `bool` and `null` as well as `Map` and `List` values. 1603 /// `String`, `bool` and `null` as well as `Map` and `List` values.
1037 */
1038 core.Map<core.String, core.Object> response; 1604 core.Map<core.String, core.Object> response;
1039 1605
1040 GoogleLongrunningOperation(); 1606 GoogleLongrunningOperation();
1041 1607
1042 GoogleLongrunningOperation.fromJson(core.Map _json) { 1608 GoogleLongrunningOperation.fromJson(core.Map _json) {
1043 if (_json.containsKey("done")) { 1609 if (_json.containsKey("done")) {
1044 done = _json["done"]; 1610 done = _json["done"];
1045 } 1611 }
1046 if (_json.containsKey("error")) { 1612 if (_json.containsKey("error")) {
1047 error = new GoogleRpcStatus.fromJson(_json["error"]); 1613 error = new GoogleRpcStatus.fromJson(_json["error"]);
1048 } 1614 }
1049 if (_json.containsKey("metadata")) { 1615 if (_json.containsKey("metadata")) {
1050 metadata = _json["metadata"]; 1616 metadata = _json["metadata"];
1051 } 1617 }
1052 if (_json.containsKey("name")) { 1618 if (_json.containsKey("name")) {
1053 name = _json["name"]; 1619 name = _json["name"];
1054 } 1620 }
1055 if (_json.containsKey("response")) { 1621 if (_json.containsKey("response")) {
1056 response = _json["response"]; 1622 response = _json["response"];
1057 } 1623 }
1058 } 1624 }
1059 1625
1060 core.Map<core.String, core.Object> toJson() { 1626 core.Map<core.String, core.Object> toJson() {
1061 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 1627 final core.Map<core.String, core.Object> _json =
1628 new core.Map<core.String, core.Object>();
1062 if (done != null) { 1629 if (done != null) {
1063 _json["done"] = done; 1630 _json["done"] = done;
1064 } 1631 }
1065 if (error != null) { 1632 if (error != null) {
1066 _json["error"] = (error).toJson(); 1633 _json["error"] = (error).toJson();
1067 } 1634 }
1068 if (metadata != null) { 1635 if (metadata != null) {
1069 _json["metadata"] = metadata; 1636 _json["metadata"] = metadata;
1070 } 1637 }
1071 if (name != null) { 1638 if (name != null) {
1072 _json["name"] = name; 1639 _json["name"] = name;
1073 } 1640 }
1074 if (response != null) { 1641 if (response != null) {
1075 _json["response"] = response; 1642 _json["response"] = response;
1076 } 1643 }
1077 return _json; 1644 return _json;
1078 } 1645 }
1079 } 1646 }
1080 1647
1081 /** 1648 /// The `Status` type defines a logical error model that is suitable for
1082 * The `Status` type defines a logical error model that is suitable for 1649 /// different
1083 * different 1650 /// programming environments, including REST APIs and RPC APIs. It is used by
1084 * programming environments, including REST APIs and RPC APIs. It is used by 1651 /// [gRPC](https://github.com/grpc). The error model is designed to be:
1085 * [gRPC](https://github.com/grpc). The error model is designed to be: 1652 ///
1086 * 1653 /// - Simple to use and understand for most users
1087 * - Simple to use and understand for most users 1654 /// - Flexible enough to meet unexpected needs
1088 * - Flexible enough to meet unexpected needs 1655 ///
1089 * 1656 /// # Overview
1090 * # Overview 1657 ///
1091 * 1658 /// The `Status` message contains three pieces of data: error code, error
1092 * The `Status` message contains three pieces of data: error code, error 1659 /// message,
1093 * message, 1660 /// and error details. The error code should be an enum value of
1094 * and error details. The error code should be an enum value of 1661 /// google.rpc.Code, but it may accept additional error codes if needed. The
1095 * google.rpc.Code, but it may accept additional error codes if needed. The 1662 /// error message should be a developer-facing English message that helps
1096 * error message should be a developer-facing English message that helps 1663 /// developers *understand* and *resolve* the error. If a localized user-facing
1097 * developers *understand* and *resolve* the error. If a localized user-facing 1664 /// error message is needed, put the localized message in the error details or
1098 * error message is needed, put the localized message in the error details or 1665 /// localize it in the client. The optional error details may contain arbitrary
1099 * localize it in the client. The optional error details may contain arbitrary 1666 /// information about the error. There is a predefined set of error detail
1100 * information about the error. There is a predefined set of error detail types 1667 /// types
1101 * in the package `google.rpc` that can be used for common error conditions. 1668 /// in the package `google.rpc` that can be used for common error conditions.
1102 * 1669 ///
1103 * # Language mapping 1670 /// # Language mapping
1104 * 1671 ///
1105 * The `Status` message is the logical representation of the error model, but it 1672 /// The `Status` message is the logical representation of the error model, but
1106 * is not necessarily the actual wire format. When the `Status` message is 1673 /// it
1107 * exposed in different client libraries and different wire protocols, it can be 1674 /// is not necessarily the actual wire format. When the `Status` message is
1108 * mapped differently. For example, it will likely be mapped to some exceptions 1675 /// exposed in different client libraries and different wire protocols, it can
1109 * in Java, but more likely mapped to some error codes in C. 1676 /// be
1110 * 1677 /// mapped differently. For example, it will likely be mapped to some
1111 * # Other uses 1678 /// exceptions
1112 * 1679 /// in Java, but more likely mapped to some error codes in C.
1113 * The error model and the `Status` message can be used in a variety of 1680 ///
1114 * environments, either with or without APIs, to provide a 1681 /// # Other uses
1115 * consistent developer experience across different environments. 1682 ///
1116 * 1683 /// The error model and the `Status` message can be used in a variety of
1117 * Example uses of this error model include: 1684 /// environments, either with or without APIs, to provide a
1118 * 1685 /// consistent developer experience across different environments.
1119 * - Partial errors. If a service needs to return partial errors to the client, 1686 ///
1120 * it may embed the `Status` in the normal response to indicate the partial 1687 /// Example uses of this error model include:
1121 * errors. 1688 ///
1122 * 1689 /// - Partial errors. If a service needs to return partial errors to the
1123 * - Workflow errors. A typical workflow has multiple steps. Each step may 1690 /// client,
1124 * have a `Status` message for error reporting. 1691 /// it may embed the `Status` in the normal response to indicate the partial
1125 * 1692 /// errors.
1126 * - Batch operations. If a client uses batch request and batch response, the 1693 ///
1127 * `Status` message should be used directly inside batch response, one for 1694 /// - Workflow errors. A typical workflow has multiple steps. Each step may
1128 * each error sub-response. 1695 /// have a `Status` message for error reporting.
1129 * 1696 ///
1130 * - Asynchronous operations. If an API call embeds asynchronous operation 1697 /// - Batch operations. If a client uses batch request and batch response, the
1131 * results in its response, the status of those operations should be 1698 /// `Status` message should be used directly inside batch response, one for
1132 * represented directly using the `Status` message. 1699 /// each error sub-response.
1133 * 1700 ///
1134 * - Logging. If some API errors are stored in logs, the message `Status` could 1701 /// - Asynchronous operations. If an API call embeds asynchronous operation
1135 * be used directly after any stripping needed for security/privacy reasons. 1702 /// results in its response, the status of those operations should be
1136 */ 1703 /// represented directly using the `Status` message.
1704 ///
1705 /// - Logging. If some API errors are stored in logs, the message `Status`
1706 /// could
1707 /// be used directly after any stripping needed for security/privacy reasons.
1137 class GoogleRpcStatus { 1708 class GoogleRpcStatus {
1138 /** The status code, which should be an enum value of google.rpc.Code. */ 1709 /// The status code, which should be an enum value of google.rpc.Code.
1139 core.int code; 1710 core.int code;
1140 /** 1711
1141 * A list of messages that carry the error details. There is a common set of 1712 /// A list of messages that carry the error details. There is a common set
1142 * message types for APIs to use. 1713 /// of
1143 * 1714 /// message types for APIs to use.
1144 * The values for Object must be JSON objects. It can consist of `num`, 1715 ///
1145 * `String`, `bool` and `null` as well as `Map` and `List` values. 1716 /// The values for Object must be JSON objects. It can consist of `num`,
1146 */ 1717 /// `String`, `bool` and `null` as well as `Map` and `List` values.
1147 core.List<core.Map<core.String, core.Object>> details; 1718 core.List<core.Map<core.String, core.Object>> details;
1148 /** 1719
1149 * A developer-facing error message, which should be in English. Any 1720 /// A developer-facing error message, which should be in English. Any
1150 * user-facing error message should be localized and sent in the 1721 /// user-facing error message should be localized and sent in the
1151 * google.rpc.Status.details field, or localized by the client. 1722 /// google.rpc.Status.details field, or localized by the client.
1152 */
1153 core.String message; 1723 core.String message;
1154 1724
1155 GoogleRpcStatus(); 1725 GoogleRpcStatus();
1156 1726
1157 GoogleRpcStatus.fromJson(core.Map _json) { 1727 GoogleRpcStatus.fromJson(core.Map _json) {
1158 if (_json.containsKey("code")) { 1728 if (_json.containsKey("code")) {
1159 code = _json["code"]; 1729 code = _json["code"];
1160 } 1730 }
1161 if (_json.containsKey("details")) { 1731 if (_json.containsKey("details")) {
1162 details = _json["details"]; 1732 details = _json["details"];
1163 } 1733 }
1164 if (_json.containsKey("message")) { 1734 if (_json.containsKey("message")) {
1165 message = _json["message"]; 1735 message = _json["message"];
1166 } 1736 }
1167 } 1737 }
1168 1738
1169 core.Map<core.String, core.Object> toJson() { 1739 core.Map<core.String, core.Object> toJson() {
1170 final core.Map<core.String, core.Object> _json = new core.Map<core.String, c ore.Object>(); 1740 final core.Map<core.String, core.Object> _json =
1741 new core.Map<core.String, core.Object>();
1171 if (code != null) { 1742 if (code != null) {
1172 _json["code"] = code; 1743 _json["code"] = code;
1173 } 1744 }
1174 if (details != null) { 1745 if (details != null) {
1175 _json["details"] = details; 1746 _json["details"] = details;
1176 } 1747 }
1177 if (message != null) { 1748 if (message != null) {
1178 _json["message"] = message; 1749 _json["message"] = message;
1179 } 1750 }
1180 return _json; 1751 return _json;
1181 } 1752 }
1182 } 1753 }
OLDNEW
« no previous file with comments | « generated/googleapis_beta/lib/toolresults/v1beta3firstparty.dart ('k') | generated/googleapis_beta/pubspec.yaml » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698