OLD | NEW |
1 // This is a generated file (see the discoveryapis_generator project). | 1 // This is a generated file (see the discoveryapis_generator project). |
2 | 2 |
3 library googleapis_beta.dataflow.v1b3; | 3 library googleapis_beta.dataflow.v1b3; |
4 | 4 |
5 import 'dart:core' as core; | 5 import 'dart:core' as core; |
6 import 'dart:async' as async; | 6 import 'dart:async' as async; |
7 import 'dart:convert' as convert; | 7 import 'dart:convert' as convert; |
8 | 8 |
9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons; | 9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons; |
10 import 'package:http/http.dart' as http; | 10 import 'package:http/http.dart' as http; |
(...skipping 90 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
101 | 101 |
102 /** | 102 /** |
103 * Creates a Cloud Dataflow job. | 103 * Creates a Cloud Dataflow job. |
104 * | 104 * |
105 * [request] - The metadata request object. | 105 * [request] - The metadata request object. |
106 * | 106 * |
107 * Request parameters: | 107 * Request parameters: |
108 * | 108 * |
109 * [projectId] - The ID of the Cloud Platform project that the job belongs to. | 109 * [projectId] - The ID of the Cloud Platform project that the job belongs to. |
110 * | 110 * |
| 111 * [location] - The location that contains this job. |
| 112 * |
| 113 * [replaceJobId] - Deprecated. This field is now in the Job message. |
| 114 * |
111 * [view] - The level of information requested in response. | 115 * [view] - The level of information requested in response. |
112 * Possible string values are: | 116 * Possible string values are: |
113 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. | 117 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. |
114 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. | 118 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. |
115 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. | 119 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. |
116 * | 120 * |
117 * [replaceJobId] - Deprecated. This field is now in the Job message. | |
118 * | |
119 * [location] - The location that contains this job. | |
120 * | |
121 * Completes with a [Job]. | 121 * Completes with a [Job]. |
122 * | 122 * |
123 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 123 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
124 * error. | 124 * error. |
125 * | 125 * |
126 * If the used [http.Client] completes with an error when making a REST call, | 126 * If the used [http.Client] completes with an error when making a REST call, |
127 * this method will complete with the same error. | 127 * this method will complete with the same error. |
128 */ | 128 */ |
129 async.Future<Job> create(Job request, core.String projectId, {core.String view
, core.String replaceJobId, core.String location}) { | 129 async.Future<Job> create(Job request, core.String projectId, {core.String loca
tion, core.String replaceJobId, core.String view}) { |
130 var _url = null; | 130 var _url = null; |
131 var _queryParams = new core.Map(); | 131 var _queryParams = new core.Map(); |
132 var _uploadMedia = null; | 132 var _uploadMedia = null; |
133 var _uploadOptions = null; | 133 var _uploadOptions = null; |
134 var _downloadOptions = commons.DownloadOptions.Metadata; | 134 var _downloadOptions = commons.DownloadOptions.Metadata; |
135 var _body = null; | 135 var _body = null; |
136 | 136 |
137 if (request != null) { | 137 if (request != null) { |
138 _body = convert.JSON.encode((request).toJson()); | 138 _body = convert.JSON.encode((request).toJson()); |
139 } | 139 } |
140 if (projectId == null) { | 140 if (projectId == null) { |
141 throw new core.ArgumentError("Parameter projectId is required."); | 141 throw new core.ArgumentError("Parameter projectId is required."); |
142 } | 142 } |
| 143 if (location != null) { |
| 144 _queryParams["location"] = [location]; |
| 145 } |
| 146 if (replaceJobId != null) { |
| 147 _queryParams["replaceJobId"] = [replaceJobId]; |
| 148 } |
143 if (view != null) { | 149 if (view != null) { |
144 _queryParams["view"] = [view]; | 150 _queryParams["view"] = [view]; |
145 } | 151 } |
146 if (replaceJobId != null) { | |
147 _queryParams["replaceJobId"] = [replaceJobId]; | |
148 } | |
149 if (location != null) { | |
150 _queryParams["location"] = [location]; | |
151 } | |
152 | 152 |
153 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs'; | 153 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs'; |
154 | 154 |
155 var _response = _requester.request(_url, | 155 var _response = _requester.request(_url, |
156 "POST", | 156 "POST", |
157 body: _body, | 157 body: _body, |
158 queryParams: _queryParams, | 158 queryParams: _queryParams, |
159 uploadOptions: _uploadOptions, | 159 uploadOptions: _uploadOptions, |
160 uploadMedia: _uploadMedia, | 160 uploadMedia: _uploadMedia, |
161 downloadOptions: _downloadOptions); | 161 downloadOptions: _downloadOptions); |
(...skipping 60 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
222 | 222 |
223 /** | 223 /** |
224 * Request the job status. | 224 * Request the job status. |
225 * | 225 * |
226 * Request parameters: | 226 * Request parameters: |
227 * | 227 * |
228 * [projectId] - A project id. | 228 * [projectId] - A project id. |
229 * | 229 * |
230 * [jobId] - The job to get messages for. | 230 * [jobId] - The job to get messages for. |
231 * | 231 * |
| 232 * [location] - The location which contains the job specified by job_id. |
| 233 * |
232 * [startTime] - Return only metric data that has changed since this time. | 234 * [startTime] - Return only metric data that has changed since this time. |
233 * Default is to return all information about all metrics for the job. | 235 * Default is to return all information about all metrics for the job. |
234 * | 236 * |
235 * [location] - The location which contains the job specified by job_id. | |
236 * | |
237 * Completes with a [JobMetrics]. | 237 * Completes with a [JobMetrics]. |
238 * | 238 * |
239 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 239 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
240 * error. | 240 * error. |
241 * | 241 * |
242 * If the used [http.Client] completes with an error when making a REST call, | 242 * If the used [http.Client] completes with an error when making a REST call, |
243 * this method will complete with the same error. | 243 * this method will complete with the same error. |
244 */ | 244 */ |
245 async.Future<JobMetrics> getMetrics(core.String projectId, core.String jobId,
{core.String startTime, core.String location}) { | 245 async.Future<JobMetrics> getMetrics(core.String projectId, core.String jobId,
{core.String location, core.String startTime}) { |
246 var _url = null; | 246 var _url = null; |
247 var _queryParams = new core.Map(); | 247 var _queryParams = new core.Map(); |
248 var _uploadMedia = null; | 248 var _uploadMedia = null; |
249 var _uploadOptions = null; | 249 var _uploadOptions = null; |
250 var _downloadOptions = commons.DownloadOptions.Metadata; | 250 var _downloadOptions = commons.DownloadOptions.Metadata; |
251 var _body = null; | 251 var _body = null; |
252 | 252 |
253 if (projectId == null) { | 253 if (projectId == null) { |
254 throw new core.ArgumentError("Parameter projectId is required."); | 254 throw new core.ArgumentError("Parameter projectId is required."); |
255 } | 255 } |
256 if (jobId == null) { | 256 if (jobId == null) { |
257 throw new core.ArgumentError("Parameter jobId is required."); | 257 throw new core.ArgumentError("Parameter jobId is required."); |
258 } | 258 } |
| 259 if (location != null) { |
| 260 _queryParams["location"] = [location]; |
| 261 } |
259 if (startTime != null) { | 262 if (startTime != null) { |
260 _queryParams["startTime"] = [startTime]; | 263 _queryParams["startTime"] = [startTime]; |
261 } | 264 } |
262 if (location != null) { | |
263 _queryParams["location"] = [location]; | |
264 } | |
265 | 265 |
266 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs/' + commons.Escaper.ecapeVariable('$jobId') + '/metrics'; | 266 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs/' + commons.Escaper.ecapeVariable('$jobId') + '/metrics'; |
267 | 267 |
268 var _response = _requester.request(_url, | 268 var _response = _requester.request(_url, |
269 "GET", | 269 "GET", |
270 body: _body, | 270 body: _body, |
271 queryParams: _queryParams, | 271 queryParams: _queryParams, |
272 uploadOptions: _uploadOptions, | 272 uploadOptions: _uploadOptions, |
273 uploadMedia: _uploadMedia, | 273 uploadMedia: _uploadMedia, |
274 downloadOptions: _downloadOptions); | 274 downloadOptions: _downloadOptions); |
275 return _response.then((data) => new JobMetrics.fromJson(data)); | 275 return _response.then((data) => new JobMetrics.fromJson(data)); |
276 } | 276 } |
277 | 277 |
278 /** | 278 /** |
279 * List the jobs of a project. | 279 * List the jobs of a project. |
280 * | 280 * |
281 * Request parameters: | 281 * Request parameters: |
282 * | 282 * |
283 * [projectId] - The project which owns the jobs. | 283 * [projectId] - The project which owns the jobs. |
284 * | 284 * |
285 * [filter] - The kind of filter to use. | 285 * [filter] - The kind of filter to use. |
286 * Possible string values are: | 286 * Possible string values are: |
287 * - "UNKNOWN" : A UNKNOWN. | 287 * - "UNKNOWN" : A UNKNOWN. |
288 * - "ALL" : A ALL. | 288 * - "ALL" : A ALL. |
289 * - "TERMINATED" : A TERMINATED. | 289 * - "TERMINATED" : A TERMINATED. |
290 * - "ACTIVE" : A ACTIVE. | 290 * - "ACTIVE" : A ACTIVE. |
291 * | 291 * |
| 292 * [location] - The location that contains this job. |
| 293 * |
| 294 * [pageToken] - Set this to the 'next_page_token' field of a previous |
| 295 * response |
| 296 * to request additional results in a long list. |
| 297 * |
| 298 * [pageSize] - If there are many jobs, limit response to at most this many. |
| 299 * The actual number of jobs returned will be the lesser of max_responses |
| 300 * and an unspecified server-defined limit. |
| 301 * |
292 * [view] - Level of information requested in response. Default is | 302 * [view] - Level of information requested in response. Default is |
293 * `JOB_VIEW_SUMMARY`. | 303 * `JOB_VIEW_SUMMARY`. |
294 * Possible string values are: | 304 * Possible string values are: |
295 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. | 305 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. |
296 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. | 306 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. |
297 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. | 307 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. |
298 * | 308 * |
299 * [pageSize] - If there are many jobs, limit response to at most this many. | |
300 * The actual number of jobs returned will be the lesser of max_responses and | |
301 * an unspecified server-defined limit. | |
302 * | |
303 * [pageToken] - Set this to the 'next_page_token' field of a previous | |
304 * response to request additional results in a long list. | |
305 * | |
306 * [location] - The location that contains this job. | |
307 * | |
308 * Completes with a [ListJobsResponse]. | 309 * Completes with a [ListJobsResponse]. |
309 * | 310 * |
310 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 311 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
311 * error. | 312 * error. |
312 * | 313 * |
313 * If the used [http.Client] completes with an error when making a REST call, | 314 * If the used [http.Client] completes with an error when making a REST call, |
314 * this method will complete with the same error. | 315 * this method will complete with the same error. |
315 */ | 316 */ |
316 async.Future<ListJobsResponse> list(core.String projectId, {core.String filter
, core.String view, core.int pageSize, core.String pageToken, core.String locati
on}) { | 317 async.Future<ListJobsResponse> list(core.String projectId, {core.String filter
, core.String location, core.String pageToken, core.int pageSize, core.String vi
ew}) { |
317 var _url = null; | 318 var _url = null; |
318 var _queryParams = new core.Map(); | 319 var _queryParams = new core.Map(); |
319 var _uploadMedia = null; | 320 var _uploadMedia = null; |
320 var _uploadOptions = null; | 321 var _uploadOptions = null; |
321 var _downloadOptions = commons.DownloadOptions.Metadata; | 322 var _downloadOptions = commons.DownloadOptions.Metadata; |
322 var _body = null; | 323 var _body = null; |
323 | 324 |
324 if (projectId == null) { | 325 if (projectId == null) { |
325 throw new core.ArgumentError("Parameter projectId is required."); | 326 throw new core.ArgumentError("Parameter projectId is required."); |
326 } | 327 } |
327 if (filter != null) { | 328 if (filter != null) { |
328 _queryParams["filter"] = [filter]; | 329 _queryParams["filter"] = [filter]; |
329 } | 330 } |
| 331 if (location != null) { |
| 332 _queryParams["location"] = [location]; |
| 333 } |
| 334 if (pageToken != null) { |
| 335 _queryParams["pageToken"] = [pageToken]; |
| 336 } |
| 337 if (pageSize != null) { |
| 338 _queryParams["pageSize"] = ["${pageSize}"]; |
| 339 } |
330 if (view != null) { | 340 if (view != null) { |
331 _queryParams["view"] = [view]; | 341 _queryParams["view"] = [view]; |
332 } | 342 } |
333 if (pageSize != null) { | |
334 _queryParams["pageSize"] = ["${pageSize}"]; | |
335 } | |
336 if (pageToken != null) { | |
337 _queryParams["pageToken"] = [pageToken]; | |
338 } | |
339 if (location != null) { | |
340 _queryParams["location"] = [location]; | |
341 } | |
342 | 343 |
343 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs'; | 344 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs'; |
344 | 345 |
345 var _response = _requester.request(_url, | 346 var _response = _requester.request(_url, |
346 "GET", | 347 "GET", |
347 body: _body, | 348 body: _body, |
348 queryParams: _queryParams, | 349 queryParams: _queryParams, |
349 uploadOptions: _uploadOptions, | 350 uploadOptions: _uploadOptions, |
350 uploadMedia: _uploadMedia, | 351 uploadMedia: _uploadMedia, |
351 downloadOptions: _downloadOptions); | 352 downloadOptions: _downloadOptions); |
(...skipping 181 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
533 * | 534 * |
534 * [minimumImportance] - Filter to only get messages with importance >= level | 535 * [minimumImportance] - Filter to only get messages with importance >= level |
535 * Possible string values are: | 536 * Possible string values are: |
536 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN. | 537 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN. |
537 * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG. | 538 * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG. |
538 * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED. | 539 * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED. |
539 * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC. | 540 * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC. |
540 * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING. | 541 * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING. |
541 * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR. | 542 * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR. |
542 * | 543 * |
543 * [pageSize] - If specified, determines the maximum number of messages to | 544 * [location] - The location which contains the job specified by job_id. |
544 * return. If unspecified, the service may choose an appropriate default, or | 545 * |
545 * may return an arbitrarily large number of results. | 546 * [endTime] - Return only messages with timestamps < end_time. The default is |
| 547 * now |
| 548 * (i.e. return up to the latest messages available). |
| 549 * |
| 550 * [startTime] - If specified, return only messages with timestamps >= |
| 551 * start_time. |
| 552 * The default is the job creation time (i.e. beginning of messages). |
546 * | 553 * |
547 * [pageToken] - If supplied, this should be the value of next_page_token | 554 * [pageToken] - If supplied, this should be the value of next_page_token |
548 * returned by an earlier call. This will cause the next page of results to be | 555 * returned |
549 * returned. | 556 * by an earlier call. This will cause the next page of results to |
| 557 * be returned. |
550 * | 558 * |
551 * [startTime] - If specified, return only messages with timestamps >= | 559 * [pageSize] - If specified, determines the maximum number of messages to |
552 * start_time. The default is the job creation time (i.e. beginning of | 560 * return. If unspecified, the service may choose an appropriate |
553 * messages). | 561 * default, or may return an arbitrarily large number of results. |
554 * | |
555 * [endTime] - Return only messages with timestamps < end_time. The default is | |
556 * now (i.e. return up to the latest messages available). | |
557 * | |
558 * [location] - The location which contains the job specified by job_id. | |
559 * | 562 * |
560 * Completes with a [ListJobMessagesResponse]. | 563 * Completes with a [ListJobMessagesResponse]. |
561 * | 564 * |
562 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 565 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
563 * error. | 566 * error. |
564 * | 567 * |
565 * If the used [http.Client] completes with an error when making a REST call, | 568 * If the used [http.Client] completes with an error when making a REST call, |
566 * this method will complete with the same error. | 569 * this method will complete with the same error. |
567 */ | 570 */ |
568 async.Future<ListJobMessagesResponse> list(core.String projectId, core.String
jobId, {core.String minimumImportance, core.int pageSize, core.String pageToken,
core.String startTime, core.String endTime, core.String location}) { | 571 async.Future<ListJobMessagesResponse> list(core.String projectId, core.String
jobId, {core.String minimumImportance, core.String location, core.String endTime
, core.String startTime, core.String pageToken, core.int pageSize}) { |
569 var _url = null; | 572 var _url = null; |
570 var _queryParams = new core.Map(); | 573 var _queryParams = new core.Map(); |
571 var _uploadMedia = null; | 574 var _uploadMedia = null; |
572 var _uploadOptions = null; | 575 var _uploadOptions = null; |
573 var _downloadOptions = commons.DownloadOptions.Metadata; | 576 var _downloadOptions = commons.DownloadOptions.Metadata; |
574 var _body = null; | 577 var _body = null; |
575 | 578 |
576 if (projectId == null) { | 579 if (projectId == null) { |
577 throw new core.ArgumentError("Parameter projectId is required."); | 580 throw new core.ArgumentError("Parameter projectId is required."); |
578 } | 581 } |
579 if (jobId == null) { | 582 if (jobId == null) { |
580 throw new core.ArgumentError("Parameter jobId is required."); | 583 throw new core.ArgumentError("Parameter jobId is required."); |
581 } | 584 } |
582 if (minimumImportance != null) { | 585 if (minimumImportance != null) { |
583 _queryParams["minimumImportance"] = [minimumImportance]; | 586 _queryParams["minimumImportance"] = [minimumImportance]; |
584 } | 587 } |
| 588 if (location != null) { |
| 589 _queryParams["location"] = [location]; |
| 590 } |
| 591 if (endTime != null) { |
| 592 _queryParams["endTime"] = [endTime]; |
| 593 } |
| 594 if (startTime != null) { |
| 595 _queryParams["startTime"] = [startTime]; |
| 596 } |
| 597 if (pageToken != null) { |
| 598 _queryParams["pageToken"] = [pageToken]; |
| 599 } |
585 if (pageSize != null) { | 600 if (pageSize != null) { |
586 _queryParams["pageSize"] = ["${pageSize}"]; | 601 _queryParams["pageSize"] = ["${pageSize}"]; |
587 } | 602 } |
588 if (pageToken != null) { | |
589 _queryParams["pageToken"] = [pageToken]; | |
590 } | |
591 if (startTime != null) { | |
592 _queryParams["startTime"] = [startTime]; | |
593 } | |
594 if (endTime != null) { | |
595 _queryParams["endTime"] = [endTime]; | |
596 } | |
597 if (location != null) { | |
598 _queryParams["location"] = [location]; | |
599 } | |
600 | 603 |
601 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs/' + commons.Escaper.ecapeVariable('$jobId') + '/messages'; | 604 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jo
bs/' + commons.Escaper.ecapeVariable('$jobId') + '/messages'; |
602 | 605 |
603 var _response = _requester.request(_url, | 606 var _response = _requester.request(_url, |
604 "GET", | 607 "GET", |
605 body: _body, | 608 body: _body, |
606 queryParams: _queryParams, | 609 queryParams: _queryParams, |
607 uploadOptions: _uploadOptions, | 610 uploadOptions: _uploadOptions, |
608 uploadMedia: _uploadMedia, | 611 uploadMedia: _uploadMedia, |
609 downloadOptions: _downloadOptions); | 612 downloadOptions: _downloadOptions); |
(...skipping 314 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
924 * | 927 * |
925 * [location] - The location that contains this job. | 928 * [location] - The location that contains this job. |
926 * | 929 * |
927 * [filter] - The kind of filter to use. | 930 * [filter] - The kind of filter to use. |
928 * Possible string values are: | 931 * Possible string values are: |
929 * - "UNKNOWN" : A UNKNOWN. | 932 * - "UNKNOWN" : A UNKNOWN. |
930 * - "ALL" : A ALL. | 933 * - "ALL" : A ALL. |
931 * - "TERMINATED" : A TERMINATED. | 934 * - "TERMINATED" : A TERMINATED. |
932 * - "ACTIVE" : A ACTIVE. | 935 * - "ACTIVE" : A ACTIVE. |
933 * | 936 * |
| 937 * [pageToken] - Set this to the 'next_page_token' field of a previous |
| 938 * response |
| 939 * to request additional results in a long list. |
| 940 * |
| 941 * [pageSize] - If there are many jobs, limit response to at most this many. |
| 942 * The actual number of jobs returned will be the lesser of max_responses |
| 943 * and an unspecified server-defined limit. |
| 944 * |
934 * [view] - Level of information requested in response. Default is | 945 * [view] - Level of information requested in response. Default is |
935 * `JOB_VIEW_SUMMARY`. | 946 * `JOB_VIEW_SUMMARY`. |
936 * Possible string values are: | 947 * Possible string values are: |
937 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. | 948 * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN. |
938 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. | 949 * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY. |
939 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. | 950 * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL. |
940 * | 951 * |
941 * [pageSize] - If there are many jobs, limit response to at most this many. | |
942 * The actual number of jobs returned will be the lesser of max_responses and | |
943 * an unspecified server-defined limit. | |
944 * | |
945 * [pageToken] - Set this to the 'next_page_token' field of a previous | |
946 * response to request additional results in a long list. | |
947 * | |
948 * Completes with a [ListJobsResponse]. | 952 * Completes with a [ListJobsResponse]. |
949 * | 953 * |
950 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 954 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
951 * error. | 955 * error. |
952 * | 956 * |
953 * If the used [http.Client] completes with an error when making a REST call, | 957 * If the used [http.Client] completes with an error when making a REST call, |
954 * this method will complete with the same error. | 958 * this method will complete with the same error. |
955 */ | 959 */ |
956 async.Future<ListJobsResponse> list(core.String projectId, core.String locatio
n, {core.String filter, core.String view, core.int pageSize, core.String pageTok
en}) { | 960 async.Future<ListJobsResponse> list(core.String projectId, core.String locatio
n, {core.String filter, core.String pageToken, core.int pageSize, core.String vi
ew}) { |
957 var _url = null; | 961 var _url = null; |
958 var _queryParams = new core.Map(); | 962 var _queryParams = new core.Map(); |
959 var _uploadMedia = null; | 963 var _uploadMedia = null; |
960 var _uploadOptions = null; | 964 var _uploadOptions = null; |
961 var _downloadOptions = commons.DownloadOptions.Metadata; | 965 var _downloadOptions = commons.DownloadOptions.Metadata; |
962 var _body = null; | 966 var _body = null; |
963 | 967 |
964 if (projectId == null) { | 968 if (projectId == null) { |
965 throw new core.ArgumentError("Parameter projectId is required."); | 969 throw new core.ArgumentError("Parameter projectId is required."); |
966 } | 970 } |
967 if (location == null) { | 971 if (location == null) { |
968 throw new core.ArgumentError("Parameter location is required."); | 972 throw new core.ArgumentError("Parameter location is required."); |
969 } | 973 } |
970 if (filter != null) { | 974 if (filter != null) { |
971 _queryParams["filter"] = [filter]; | 975 _queryParams["filter"] = [filter]; |
972 } | 976 } |
| 977 if (pageToken != null) { |
| 978 _queryParams["pageToken"] = [pageToken]; |
| 979 } |
| 980 if (pageSize != null) { |
| 981 _queryParams["pageSize"] = ["${pageSize}"]; |
| 982 } |
973 if (view != null) { | 983 if (view != null) { |
974 _queryParams["view"] = [view]; | 984 _queryParams["view"] = [view]; |
975 } | 985 } |
976 if (pageSize != null) { | |
977 _queryParams["pageSize"] = ["${pageSize}"]; | |
978 } | |
979 if (pageToken != null) { | |
980 _queryParams["pageToken"] = [pageToken]; | |
981 } | |
982 | 986 |
983 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/lo
cations/' + commons.Escaper.ecapeVariable('$location') + '/jobs'; | 987 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/lo
cations/' + commons.Escaper.ecapeVariable('$location') + '/jobs'; |
984 | 988 |
985 var _response = _requester.request(_url, | 989 var _response = _requester.request(_url, |
986 "GET", | 990 "GET", |
987 body: _body, | 991 body: _body, |
988 queryParams: _queryParams, | 992 queryParams: _queryParams, |
989 uploadOptions: _uploadOptions, | 993 uploadOptions: _uploadOptions, |
990 uploadMedia: _uploadMedia, | 994 uploadMedia: _uploadMedia, |
991 downloadOptions: _downloadOptions); | 995 downloadOptions: _downloadOptions); |
(...skipping 67 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1059 * Request the job status. | 1063 * Request the job status. |
1060 * | 1064 * |
1061 * Request parameters: | 1065 * Request parameters: |
1062 * | 1066 * |
1063 * [projectId] - A project id. | 1067 * [projectId] - A project id. |
1064 * | 1068 * |
1065 * [location] - The location which contains the job specified by job_id. | 1069 * [location] - The location which contains the job specified by job_id. |
1066 * | 1070 * |
1067 * [jobId] - The job to get messages about. | 1071 * [jobId] - The job to get messages about. |
1068 * | 1072 * |
| 1073 * [endTime] - Return only messages with timestamps < end_time. The default is |
| 1074 * now |
| 1075 * (i.e. return up to the latest messages available). |
| 1076 * |
| 1077 * [startTime] - If specified, return only messages with timestamps >= |
| 1078 * start_time. |
| 1079 * The default is the job creation time (i.e. beginning of messages). |
| 1080 * |
| 1081 * [pageToken] - If supplied, this should be the value of next_page_token |
| 1082 * returned |
| 1083 * by an earlier call. This will cause the next page of results to |
| 1084 * be returned. |
| 1085 * |
| 1086 * [pageSize] - If specified, determines the maximum number of messages to |
| 1087 * return. If unspecified, the service may choose an appropriate |
| 1088 * default, or may return an arbitrarily large number of results. |
| 1089 * |
1069 * [minimumImportance] - Filter to only get messages with importance >= level | 1090 * [minimumImportance] - Filter to only get messages with importance >= level |
1070 * Possible string values are: | 1091 * Possible string values are: |
1071 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN. | 1092 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN. |
1072 * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG. | 1093 * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG. |
1073 * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED. | 1094 * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED. |
1074 * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC. | 1095 * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC. |
1075 * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING. | 1096 * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING. |
1076 * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR. | 1097 * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR. |
1077 * | 1098 * |
1078 * [pageSize] - If specified, determines the maximum number of messages to | |
1079 * return. If unspecified, the service may choose an appropriate default, or | |
1080 * may return an arbitrarily large number of results. | |
1081 * | |
1082 * [pageToken] - If supplied, this should be the value of next_page_token | |
1083 * returned by an earlier call. This will cause the next page of results to be | |
1084 * returned. | |
1085 * | |
1086 * [startTime] - If specified, return only messages with timestamps >= | |
1087 * start_time. The default is the job creation time (i.e. beginning of | |
1088 * messages). | |
1089 * | |
1090 * [endTime] - Return only messages with timestamps < end_time. The default is | |
1091 * now (i.e. return up to the latest messages available). | |
1092 * | |
1093 * Completes with a [ListJobMessagesResponse]. | 1099 * Completes with a [ListJobMessagesResponse]. |
1094 * | 1100 * |
1095 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 1101 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
1096 * error. | 1102 * error. |
1097 * | 1103 * |
1098 * If the used [http.Client] completes with an error when making a REST call, | 1104 * If the used [http.Client] completes with an error when making a REST call, |
1099 * this method will complete with the same error. | 1105 * this method will complete with the same error. |
1100 */ | 1106 */ |
1101 async.Future<ListJobMessagesResponse> list(core.String projectId, core.String
location, core.String jobId, {core.String minimumImportance, core.int pageSize,
core.String pageToken, core.String startTime, core.String endTime}) { | 1107 async.Future<ListJobMessagesResponse> list(core.String projectId, core.String
location, core.String jobId, {core.String endTime, core.String startTime, core.S
tring pageToken, core.int pageSize, core.String minimumImportance}) { |
1102 var _url = null; | 1108 var _url = null; |
1103 var _queryParams = new core.Map(); | 1109 var _queryParams = new core.Map(); |
1104 var _uploadMedia = null; | 1110 var _uploadMedia = null; |
1105 var _uploadOptions = null; | 1111 var _uploadOptions = null; |
1106 var _downloadOptions = commons.DownloadOptions.Metadata; | 1112 var _downloadOptions = commons.DownloadOptions.Metadata; |
1107 var _body = null; | 1113 var _body = null; |
1108 | 1114 |
1109 if (projectId == null) { | 1115 if (projectId == null) { |
1110 throw new core.ArgumentError("Parameter projectId is required."); | 1116 throw new core.ArgumentError("Parameter projectId is required."); |
1111 } | 1117 } |
1112 if (location == null) { | 1118 if (location == null) { |
1113 throw new core.ArgumentError("Parameter location is required."); | 1119 throw new core.ArgumentError("Parameter location is required."); |
1114 } | 1120 } |
1115 if (jobId == null) { | 1121 if (jobId == null) { |
1116 throw new core.ArgumentError("Parameter jobId is required."); | 1122 throw new core.ArgumentError("Parameter jobId is required."); |
1117 } | 1123 } |
| 1124 if (endTime != null) { |
| 1125 _queryParams["endTime"] = [endTime]; |
| 1126 } |
| 1127 if (startTime != null) { |
| 1128 _queryParams["startTime"] = [startTime]; |
| 1129 } |
| 1130 if (pageToken != null) { |
| 1131 _queryParams["pageToken"] = [pageToken]; |
| 1132 } |
| 1133 if (pageSize != null) { |
| 1134 _queryParams["pageSize"] = ["${pageSize}"]; |
| 1135 } |
1118 if (minimumImportance != null) { | 1136 if (minimumImportance != null) { |
1119 _queryParams["minimumImportance"] = [minimumImportance]; | 1137 _queryParams["minimumImportance"] = [minimumImportance]; |
1120 } | 1138 } |
1121 if (pageSize != null) { | |
1122 _queryParams["pageSize"] = ["${pageSize}"]; | |
1123 } | |
1124 if (pageToken != null) { | |
1125 _queryParams["pageToken"] = [pageToken]; | |
1126 } | |
1127 if (startTime != null) { | |
1128 _queryParams["startTime"] = [startTime]; | |
1129 } | |
1130 if (endTime != null) { | |
1131 _queryParams["endTime"] = [endTime]; | |
1132 } | |
1133 | 1139 |
1134 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/lo
cations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Esca
per.ecapeVariable('$jobId') + '/messages'; | 1140 _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/lo
cations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Esca
per.ecapeVariable('$jobId') + '/messages'; |
1135 | 1141 |
1136 var _response = _requester.request(_url, | 1142 var _response = _requester.request(_url, |
1137 "GET", | 1143 "GET", |
1138 body: _body, | 1144 body: _body, |
1139 queryParams: _queryParams, | 1145 queryParams: _queryParams, |
1140 uploadOptions: _uploadOptions, | 1146 uploadOptions: _uploadOptions, |
1141 uploadMedia: _uploadMedia, | 1147 uploadMedia: _uploadMedia, |
1142 downloadOptions: _downloadOptions); | 1148 downloadOptions: _downloadOptions); |
(...skipping 214 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1357 return _json; | 1363 return _json; |
1358 } | 1364 } |
1359 } | 1365 } |
1360 | 1366 |
1361 /** A progress measurement of a WorkItem by a worker. */ | 1367 /** A progress measurement of a WorkItem by a worker. */ |
1362 class ApproximateReportedProgress { | 1368 class ApproximateReportedProgress { |
1363 /** | 1369 /** |
1364 * Total amount of parallelism in the portion of input of this task that has | 1370 * Total amount of parallelism in the portion of input of this task that has |
1365 * already been consumed and is no longer active. In the first two examples | 1371 * already been consumed and is no longer active. In the first two examples |
1366 * above (see remaining_parallelism), the value should be 29 or 2 | 1372 * above (see remaining_parallelism), the value should be 29 or 2 |
1367 * respectively. The sum of remaining_parallelism and consumed_parallelism | 1373 * respectively. The sum of remaining_parallelism and consumed_parallelism |
1368 * should equal the total amount of parallelism in this work item. If | 1374 * should equal the total amount of parallelism in this work item. If |
1369 * specified, must be finite. | 1375 * specified, must be finite. |
1370 */ | 1376 */ |
1371 ReportedParallelism consumedParallelism; | 1377 ReportedParallelism consumedParallelism; |
1372 /** | 1378 /** |
1373 * Completion as fraction of the input consumed, from 0.0 (beginning, nothing | 1379 * Completion as fraction of the input consumed, from 0.0 (beginning, nothing |
1374 * consumed), to 1.0 (end of the input, entire input consumed). | 1380 * consumed), to 1.0 (end of the input, entire input consumed). |
1375 */ | 1381 */ |
1376 core.double fractionConsumed; | 1382 core.double fractionConsumed; |
1377 /** A Position within the work to represent a progress. */ | 1383 /** A Position within the work to represent a progress. */ |
1378 Position position; | 1384 Position position; |
1379 /** | 1385 /** |
1380 * Total amount of parallelism in the input of this task that remains, (i.e. | 1386 * Total amount of parallelism in the input of this task that remains, |
1381 * can be delegated to this task and any new tasks via dynamic splitting). | 1387 * (i.e. can be delegated to this task and any new tasks via dynamic |
1382 * Always at least 1 for non-finished work items and 0 for finished. "Amount | 1388 * splitting). Always at least 1 for non-finished work items and 0 for |
1383 * of parallelism" refers to how many non-empty parts of the input can be read | 1389 * finished. |
1384 * in parallel. This does not necessarily equal number of records. An input | 1390 * |
1385 * that can be read in parallel down to the individual records is called | 1391 * "Amount of parallelism" refers to how many non-empty parts of the input |
1386 * "perfectly splittable". An example of non-perfectly parallelizable input is | 1392 * can be read in parallel. This does not necessarily equal number |
1387 * a block-compressed file format where a block of records has to be read as a | 1393 * of records. An input that can be read in parallel down to the |
1388 * whole, but different blocks can be read in parallel. Examples: * If we are | 1394 * individual records is called "perfectly splittable". |
1389 * processing record #30 (starting at 1) out of 50 in a perfectly splittable | 1395 * An example of non-perfectly parallelizable input is a block-compressed |
1390 * 50-record input, this value should be 21 (20 remaining + 1 current). * If | 1396 * file format where a block of records has to be read as a whole, |
1391 * we are reading through block 3 in a block-compressed file consisting of 5 | 1397 * but different blocks can be read in parallel. |
1392 * blocks, this value should be 3 (since blocks 4 and 5 can be processed in | 1398 * |
1393 * parallel by new tasks via dynamic splitting and the current task remains | 1399 * Examples: |
1394 * processing block 3). * If we are reading through the last block in a | 1400 * * If we are processing record #30 (starting at 1) out of 50 in a perfectly |
1395 * block-compressed file, or reading or processing the last record in a | 1401 * splittable 50-record input, this value should be 21 (20 remaining + 1 |
1396 * perfectly splittable input, this value should be 1, because apart from the | 1402 * current). |
1397 * current task, no additional remainder can be split off. | 1403 * * If we are reading through block 3 in a block-compressed file consisting |
| 1404 * of 5 blocks, this value should be 3 (since blocks 4 and 5 can be |
| 1405 * processed in parallel by new tasks via dynamic splitting and the current |
| 1406 * task remains processing block 3). |
| 1407 * * If we are reading through the last block in a block-compressed file, |
| 1408 * or reading or processing the last record in a perfectly splittable |
| 1409 * input, this value should be 1, because apart from the current task, no |
| 1410 * additional remainder can be split off. |
1398 */ | 1411 */ |
1399 ReportedParallelism remainingParallelism; | 1412 ReportedParallelism remainingParallelism; |
1400 | 1413 |
1401 ApproximateReportedProgress(); | 1414 ApproximateReportedProgress(); |
1402 | 1415 |
1403 ApproximateReportedProgress.fromJson(core.Map _json) { | 1416 ApproximateReportedProgress.fromJson(core.Map _json) { |
1404 if (_json.containsKey("consumedParallelism")) { | 1417 if (_json.containsKey("consumedParallelism")) { |
1405 consumedParallelism = new ReportedParallelism.fromJson(_json["consumedPara
llelism"]); | 1418 consumedParallelism = new ReportedParallelism.fromJson(_json["consumedPara
llelism"]); |
1406 } | 1419 } |
1407 if (_json.containsKey("fractionConsumed")) { | 1420 if (_json.containsKey("fractionConsumed")) { |
(...skipping 58 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1466 } | 1479 } |
1467 return _json; | 1480 return _json; |
1468 } | 1481 } |
1469 } | 1482 } |
1470 | 1483 |
1471 /** Settings for WorkerPool autoscaling. */ | 1484 /** Settings for WorkerPool autoscaling. */ |
1472 class AutoscalingSettings { | 1485 class AutoscalingSettings { |
1473 /** | 1486 /** |
1474 * The algorithm to use for autoscaling. | 1487 * The algorithm to use for autoscaling. |
1475 * Possible string values are: | 1488 * Possible string values are: |
1476 * - "AUTOSCALING_ALGORITHM_UNKNOWN" : A AUTOSCALING_ALGORITHM_UNKNOWN. | 1489 * - "AUTOSCALING_ALGORITHM_UNKNOWN" : The algorithm is unknown, or |
1477 * - "AUTOSCALING_ALGORITHM_NONE" : A AUTOSCALING_ALGORITHM_NONE. | 1490 * unspecified. |
1478 * - "AUTOSCALING_ALGORITHM_BASIC" : A AUTOSCALING_ALGORITHM_BASIC. | 1491 * - "AUTOSCALING_ALGORITHM_NONE" : Disable autoscaling. |
| 1492 * - "AUTOSCALING_ALGORITHM_BASIC" : Increase worker count over time to reduce |
| 1493 * job execution time. |
1479 */ | 1494 */ |
1480 core.String algorithm; | 1495 core.String algorithm; |
1481 /** The maximum number of workers to cap scaling at. */ | 1496 /** The maximum number of workers to cap scaling at. */ |
1482 core.int maxNumWorkers; | 1497 core.int maxNumWorkers; |
1483 | 1498 |
1484 AutoscalingSettings(); | 1499 AutoscalingSettings(); |
1485 | 1500 |
1486 AutoscalingSettings.fromJson(core.Map _json) { | 1501 AutoscalingSettings.fromJson(core.Map _json) { |
1487 if (_json.containsKey("algorithm")) { | 1502 if (_json.containsKey("algorithm")) { |
1488 algorithm = _json["algorithm"]; | 1503 algorithm = _json["algorithm"]; |
1489 } | 1504 } |
1490 if (_json.containsKey("maxNumWorkers")) { | 1505 if (_json.containsKey("maxNumWorkers")) { |
1491 maxNumWorkers = _json["maxNumWorkers"]; | 1506 maxNumWorkers = _json["maxNumWorkers"]; |
1492 } | 1507 } |
1493 } | 1508 } |
1494 | 1509 |
1495 core.Map toJson() { | 1510 core.Map toJson() { |
1496 var _json = new core.Map(); | 1511 var _json = new core.Map(); |
1497 if (algorithm != null) { | 1512 if (algorithm != null) { |
1498 _json["algorithm"] = algorithm; | 1513 _json["algorithm"] = algorithm; |
1499 } | 1514 } |
1500 if (maxNumWorkers != null) { | 1515 if (maxNumWorkers != null) { |
1501 _json["maxNumWorkers"] = maxNumWorkers; | 1516 _json["maxNumWorkers"] = maxNumWorkers; |
1502 } | 1517 } |
1503 return _json; | 1518 return _json; |
1504 } | 1519 } |
1505 } | 1520 } |
1506 | 1521 |
| 1522 /** Modeled after information exposed by /proc/stat. */ |
| 1523 class CPUTime { |
| 1524 /** |
| 1525 * Average CPU utilization rate (% non-idle cpu / second) since previous |
| 1526 * sample. |
| 1527 */ |
| 1528 core.double rate; |
| 1529 /** Timestamp of the measurement. */ |
| 1530 core.String timestamp; |
| 1531 /** |
| 1532 * Total active CPU time across all cores (ie., non-idle) in milliseconds |
| 1533 * since start-up. |
| 1534 */ |
| 1535 core.String totalMs; |
| 1536 |
| 1537 CPUTime(); |
| 1538 |
| 1539 CPUTime.fromJson(core.Map _json) { |
| 1540 if (_json.containsKey("rate")) { |
| 1541 rate = _json["rate"]; |
| 1542 } |
| 1543 if (_json.containsKey("timestamp")) { |
| 1544 timestamp = _json["timestamp"]; |
| 1545 } |
| 1546 if (_json.containsKey("totalMs")) { |
| 1547 totalMs = _json["totalMs"]; |
| 1548 } |
| 1549 } |
| 1550 |
| 1551 core.Map toJson() { |
| 1552 var _json = new core.Map(); |
| 1553 if (rate != null) { |
| 1554 _json["rate"] = rate; |
| 1555 } |
| 1556 if (timestamp != null) { |
| 1557 _json["timestamp"] = timestamp; |
| 1558 } |
| 1559 if (totalMs != null) { |
| 1560 _json["totalMs"] = totalMs; |
| 1561 } |
| 1562 return _json; |
| 1563 } |
| 1564 } |
| 1565 |
1507 /** All configuration data for a particular Computation. */ | 1566 /** All configuration data for a particular Computation. */ |
1508 class ComputationTopology { | 1567 class ComputationTopology { |
1509 /** The ID of the computation. */ | 1568 /** The ID of the computation. */ |
1510 core.String computationId; | 1569 core.String computationId; |
1511 /** The inputs to the computation. */ | 1570 /** The inputs to the computation. */ |
1512 core.List<StreamLocation> inputs; | 1571 core.List<StreamLocation> inputs; |
1513 /** The key ranges processed by the computation. */ | 1572 /** The key ranges processed by the computation. */ |
1514 core.List<KeyRangeLocation> keyRanges; | 1573 core.List<KeyRangeLocation> keyRanges; |
1515 /** The outputs from the computation. */ | 1574 /** The outputs from the computation. */ |
1516 core.List<StreamLocation> outputs; | 1575 core.List<StreamLocation> outputs; |
(...skipping 93 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1610 | 1669 |
1611 /** | 1670 /** |
1612 * CounterMetadata includes all static non-name non-value counter attributes. | 1671 * CounterMetadata includes all static non-name non-value counter attributes. |
1613 */ | 1672 */ |
1614 class CounterMetadata { | 1673 class CounterMetadata { |
1615 /** Human-readable description of the counter semantics. */ | 1674 /** Human-readable description of the counter semantics. */ |
1616 core.String description; | 1675 core.String description; |
1617 /** | 1676 /** |
1618 * Counter aggregation kind. | 1677 * Counter aggregation kind. |
1619 * Possible string values are: | 1678 * Possible string values are: |
1620 * - "INVALID" : A INVALID. | 1679 * - "INVALID" : Counter aggregation kind was not set. |
1621 * - "SUM" : A SUM. | 1680 * - "SUM" : Aggregated value is the sum of all contributed values. |
1622 * - "MAX" : A MAX. | 1681 * - "MAX" : Aggregated value is the max of all contributed values. |
1623 * - "MIN" : A MIN. | 1682 * - "MIN" : Aggregated value is the min of all contributed values. |
1624 * - "MEAN" : A MEAN. | 1683 * - "MEAN" : Aggregated value is the mean of all contributed values. |
1625 * - "OR" : A OR. | 1684 * - "OR" : Aggregated value represents the logical 'or' of all contributed |
1626 * - "AND" : A AND. | 1685 * values. |
1627 * - "SET" : A SET. | 1686 * - "AND" : Aggregated value represents the logical 'and' of all contributed |
1628 * - "DISTRIBUTION" : A DISTRIBUTION. | 1687 * values. |
| 1688 * - "SET" : Aggregated value is a set of unique contributed values. |
| 1689 * - "DISTRIBUTION" : Aggregated value captures statistics about a |
| 1690 * distribution. |
1629 */ | 1691 */ |
1630 core.String kind; | 1692 core.String kind; |
1631 /** A string referring to the unit type. */ | 1693 /** A string referring to the unit type. */ |
1632 core.String otherUnits; | 1694 core.String otherUnits; |
1633 /** | 1695 /** |
1634 * System defined Units, see above enum. | 1696 * System defined Units, see above enum. |
1635 * Possible string values are: | 1697 * Possible string values are: |
1636 * - "BYTES" : A BYTES. | 1698 * - "BYTES" : Counter returns a value in bytes. |
1637 * - "BYTES_PER_SEC" : A BYTES_PER_SEC. | 1699 * - "BYTES_PER_SEC" : Counter returns a value in bytes per second. |
1638 * - "MILLISECONDS" : A MILLISECONDS. | 1700 * - "MILLISECONDS" : Counter returns a value in milliseconds. |
1639 * - "MICROSECONDS" : A MICROSECONDS. | 1701 * - "MICROSECONDS" : Counter returns a value in microseconds. |
1640 * - "NANOSECONDS" : A NANOSECONDS. | 1702 * - "NANOSECONDS" : Counter returns a value in nanoseconds. |
1641 * - "TIMESTAMP_MSEC" : A TIMESTAMP_MSEC. | 1703 * - "TIMESTAMP_MSEC" : Counter returns a timestamp in milliseconds. |
1642 * - "TIMESTAMP_USEC" : A TIMESTAMP_USEC. | 1704 * - "TIMESTAMP_USEC" : Counter returns a timestamp in microseconds. |
1643 * - "TIMESTAMP_NSEC" : A TIMESTAMP_NSEC. | 1705 * - "TIMESTAMP_NSEC" : Counter returns a timestamp in nanoseconds. |
1644 */ | 1706 */ |
1645 core.String standardUnits; | 1707 core.String standardUnits; |
1646 | 1708 |
1647 CounterMetadata(); | 1709 CounterMetadata(); |
1648 | 1710 |
1649 CounterMetadata.fromJson(core.Map _json) { | 1711 CounterMetadata.fromJson(core.Map _json) { |
1650 if (_json.containsKey("description")) { | 1712 if (_json.containsKey("description")) { |
1651 description = _json["description"]; | 1713 description = _json["description"]; |
1652 } | 1714 } |
1653 if (_json.containsKey("kind")) { | 1715 if (_json.containsKey("kind")) { |
(...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1685 */ | 1747 */ |
1686 class CounterStructuredName { | 1748 class CounterStructuredName { |
1687 /** Name of the optimized step being executed by the workers. */ | 1749 /** Name of the optimized step being executed by the workers. */ |
1688 core.String componentStepName; | 1750 core.String componentStepName; |
1689 /** | 1751 /** |
1690 * Name of the stage. An execution step contains multiple component steps. | 1752 * Name of the stage. An execution step contains multiple component steps. |
1691 */ | 1753 */ |
1692 core.String executionStepName; | 1754 core.String executionStepName; |
1693 /** | 1755 /** |
1694 * Counter name. Not necessarily globally-unique, but unique within the | 1756 * Counter name. Not necessarily globally-unique, but unique within the |
1695 * context of the other fields. Required. | 1757 * context of the other fields. |
| 1758 * Required. |
1696 */ | 1759 */ |
1697 core.String name; | 1760 core.String name; |
1698 /** | 1761 /** |
| 1762 * One of the standard Origins defined above. |
| 1763 * Possible string values are: |
| 1764 * - "SYSTEM" : Counter was created by the Dataflow system. |
| 1765 * - "USER" : Counter was created by the user. |
| 1766 */ |
| 1767 core.String origin; |
| 1768 /** A string containing a more specific namespace of the counter's origin. */ |
| 1769 core.String originNamespace; |
| 1770 /** |
1699 * System generated name of the original step in the user's graph, before | 1771 * System generated name of the original step in the user's graph, before |
1700 * optimization. | 1772 * optimization. |
1701 */ | 1773 */ |
1702 core.String originalStepName; | 1774 core.String originalStepName; |
1703 /** A string containing the origin of the counter. */ | |
1704 core.String otherOrigin; | |
1705 /** | 1775 /** |
1706 * Portion of this counter, either key or value. | 1776 * Portion of this counter, either key or value. |
1707 * Possible string values are: | 1777 * Possible string values are: |
1708 * - "ALL" : A ALL. | 1778 * - "ALL" : Counter portion has not been set. |
1709 * - "KEY" : A KEY. | 1779 * - "KEY" : Counter reports a key. |
1710 * - "VALUE" : A VALUE. | 1780 * - "VALUE" : Counter reports a value. |
1711 */ | 1781 */ |
1712 core.String portion; | 1782 core.String portion; |
1713 /** | |
1714 * One of the standard Origins defined above. | |
1715 * Possible string values are: | |
1716 * - "DATAFLOW" : A DATAFLOW. | |
1717 * - "USER" : A USER. | |
1718 */ | |
1719 core.String standardOrigin; | |
1720 /** ID of a particular worker. */ | 1783 /** ID of a particular worker. */ |
1721 core.String workerId; | 1784 core.String workerId; |
1722 | 1785 |
1723 CounterStructuredName(); | 1786 CounterStructuredName(); |
1724 | 1787 |
1725 CounterStructuredName.fromJson(core.Map _json) { | 1788 CounterStructuredName.fromJson(core.Map _json) { |
1726 if (_json.containsKey("componentStepName")) { | 1789 if (_json.containsKey("componentStepName")) { |
1727 componentStepName = _json["componentStepName"]; | 1790 componentStepName = _json["componentStepName"]; |
1728 } | 1791 } |
1729 if (_json.containsKey("executionStepName")) { | 1792 if (_json.containsKey("executionStepName")) { |
1730 executionStepName = _json["executionStepName"]; | 1793 executionStepName = _json["executionStepName"]; |
1731 } | 1794 } |
1732 if (_json.containsKey("name")) { | 1795 if (_json.containsKey("name")) { |
1733 name = _json["name"]; | 1796 name = _json["name"]; |
1734 } | 1797 } |
| 1798 if (_json.containsKey("origin")) { |
| 1799 origin = _json["origin"]; |
| 1800 } |
| 1801 if (_json.containsKey("originNamespace")) { |
| 1802 originNamespace = _json["originNamespace"]; |
| 1803 } |
1735 if (_json.containsKey("originalStepName")) { | 1804 if (_json.containsKey("originalStepName")) { |
1736 originalStepName = _json["originalStepName"]; | 1805 originalStepName = _json["originalStepName"]; |
1737 } | 1806 } |
1738 if (_json.containsKey("otherOrigin")) { | |
1739 otherOrigin = _json["otherOrigin"]; | |
1740 } | |
1741 if (_json.containsKey("portion")) { | 1807 if (_json.containsKey("portion")) { |
1742 portion = _json["portion"]; | 1808 portion = _json["portion"]; |
1743 } | 1809 } |
1744 if (_json.containsKey("standardOrigin")) { | |
1745 standardOrigin = _json["standardOrigin"]; | |
1746 } | |
1747 if (_json.containsKey("workerId")) { | 1810 if (_json.containsKey("workerId")) { |
1748 workerId = _json["workerId"]; | 1811 workerId = _json["workerId"]; |
1749 } | 1812 } |
1750 } | 1813 } |
1751 | 1814 |
1752 core.Map toJson() { | 1815 core.Map toJson() { |
1753 var _json = new core.Map(); | 1816 var _json = new core.Map(); |
1754 if (componentStepName != null) { | 1817 if (componentStepName != null) { |
1755 _json["componentStepName"] = componentStepName; | 1818 _json["componentStepName"] = componentStepName; |
1756 } | 1819 } |
1757 if (executionStepName != null) { | 1820 if (executionStepName != null) { |
1758 _json["executionStepName"] = executionStepName; | 1821 _json["executionStepName"] = executionStepName; |
1759 } | 1822 } |
1760 if (name != null) { | 1823 if (name != null) { |
1761 _json["name"] = name; | 1824 _json["name"] = name; |
1762 } | 1825 } |
| 1826 if (origin != null) { |
| 1827 _json["origin"] = origin; |
| 1828 } |
| 1829 if (originNamespace != null) { |
| 1830 _json["originNamespace"] = originNamespace; |
| 1831 } |
1763 if (originalStepName != null) { | 1832 if (originalStepName != null) { |
1764 _json["originalStepName"] = originalStepName; | 1833 _json["originalStepName"] = originalStepName; |
1765 } | 1834 } |
1766 if (otherOrigin != null) { | |
1767 _json["otherOrigin"] = otherOrigin; | |
1768 } | |
1769 if (portion != null) { | 1835 if (portion != null) { |
1770 _json["portion"] = portion; | 1836 _json["portion"] = portion; |
1771 } | 1837 } |
1772 if (standardOrigin != null) { | |
1773 _json["standardOrigin"] = standardOrigin; | |
1774 } | |
1775 if (workerId != null) { | 1838 if (workerId != null) { |
1776 _json["workerId"] = workerId; | 1839 _json["workerId"] = workerId; |
1777 } | 1840 } |
1778 return _json; | 1841 return _json; |
1779 } | 1842 } |
1780 } | 1843 } |
1781 | 1844 |
1782 /** | 1845 /** |
1783 * A single message which encapsulates structured name and metadata for a given | 1846 * A single message which encapsulates structured name and metadata for a given |
1784 * counter. | 1847 * counter. |
(...skipping 25 matching lines...) Expand all Loading... |
1810 } | 1873 } |
1811 return _json; | 1874 return _json; |
1812 } | 1875 } |
1813 } | 1876 } |
1814 | 1877 |
1815 /** An update to a Counter sent from a worker. */ | 1878 /** An update to a Counter sent from a worker. */ |
1816 class CounterUpdate { | 1879 class CounterUpdate { |
1817 /** Boolean value for And, Or. */ | 1880 /** Boolean value for And, Or. */ |
1818 core.bool boolean; | 1881 core.bool boolean; |
1819 /** | 1882 /** |
1820 * True if this counter is reported as the total cumulative aggregate value | 1883 * True if this counter is reported as the total cumulative aggregate |
1821 * accumulated since the worker started working on this WorkItem. By default | 1884 * value accumulated since the worker started working on this WorkItem. |
1822 * this is false, indicating that this counter is reported as a delta. | 1885 * By default this is false, indicating that this counter is reported |
| 1886 * as a delta. |
1823 */ | 1887 */ |
1824 core.bool cumulative; | 1888 core.bool cumulative; |
1825 /** Distribution data */ | 1889 /** Distribution data */ |
1826 DistributionUpdate distribution; | 1890 DistributionUpdate distribution; |
1827 /** Floating point value for Sum, Max, Min. */ | 1891 /** Floating point value for Sum, Max, Min. */ |
1828 core.double floatingPoint; | 1892 core.double floatingPoint; |
1829 /** List of floating point numbers, for Set. */ | 1893 /** List of floating point numbers, for Set. */ |
1830 FloatingPointList floatingPointList; | 1894 FloatingPointList floatingPointList; |
1831 /** Floating point mean aggregation value for Mean. */ | 1895 /** Floating point mean aggregation value for Mean. */ |
1832 FloatingPointMean floatingPointMean; | 1896 FloatingPointMean floatingPointMean; |
1833 /** Integer value for Sum, Max, Min. */ | 1897 /** Integer value for Sum, Max, Min. */ |
1834 SplitInt64 integer; | 1898 SplitInt64 integer; |
1835 /** List of integers, for Set. */ | 1899 /** List of integers, for Set. */ |
1836 IntegerList integerList; | 1900 IntegerList integerList; |
1837 /** Integer mean aggregation value for Mean. */ | 1901 /** Integer mean aggregation value for Mean. */ |
1838 IntegerMean integerMean; | 1902 IntegerMean integerMean; |
1839 /** | 1903 /** |
1840 * Value for internally-defined counters used by the Dataflow service. | 1904 * Value for internally-defined counters used by the Dataflow service. |
1841 * | 1905 * |
1842 * The values for Object must be JSON objects. It can consist of `num`, | 1906 * The values for Object must be JSON objects. It can consist of `num`, |
1843 * `String`, `bool` and `null` as well as `Map` and `List` values. | 1907 * `String`, `bool` and `null` as well as `Map` and `List` values. |
1844 */ | 1908 */ |
1845 core.Object internal; | 1909 core.Object internal; |
1846 /** Counter name and aggregation type. */ | 1910 /** Counter name and aggregation type. */ |
1847 NameAndKind nameAndKind; | 1911 NameAndKind nameAndKind; |
1848 /** | 1912 /** |
1849 * The service-generated short identifier for this counter. The short_id -> | 1913 * The service-generated short identifier for this counter. |
1850 * (name, metadata) mapping is constant for the lifetime of a job. | 1914 * The short_id -> (name, metadata) mapping is constant for the lifetime of |
| 1915 * a job. |
1851 */ | 1916 */ |
1852 core.String shortId; | 1917 core.String shortId; |
1853 /** List of strings, for Set. */ | 1918 /** List of strings, for Set. */ |
1854 StringList stringList; | 1919 StringList stringList; |
1855 /** Counter structured name and metadata. */ | 1920 /** Counter structured name and metadata. */ |
1856 CounterStructuredNameAndMetadata structuredNameAndMetadata; | 1921 CounterStructuredNameAndMetadata structuredNameAndMetadata; |
1857 | 1922 |
1858 CounterUpdate(); | 1923 CounterUpdate(); |
1859 | 1924 |
1860 CounterUpdate.fromJson(core.Map _json) { | 1925 CounterUpdate.fromJson(core.Map _json) { |
(...skipping 87 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1948 } | 2013 } |
1949 return _json; | 2014 return _json; |
1950 } | 2015 } |
1951 } | 2016 } |
1952 | 2017 |
1953 /** A request to create a Cloud Dataflow job from a template. */ | 2018 /** A request to create a Cloud Dataflow job from a template. */ |
1954 class CreateJobFromTemplateRequest { | 2019 class CreateJobFromTemplateRequest { |
1955 /** The runtime environment for the job. */ | 2020 /** The runtime environment for the job. */ |
1956 RuntimeEnvironment environment; | 2021 RuntimeEnvironment environment; |
1957 /** | 2022 /** |
1958 * Required. A Cloud Storage path to the template from which to create the | 2023 * Required. A Cloud Storage path to the template from which to |
1959 * job. Must be a valid Cloud Storage URL, beginning with `gs://`. | 2024 * create the job. |
| 2025 * Must be a valid Cloud Storage URL, beginning with `gs://`. |
1960 */ | 2026 */ |
1961 core.String gcsPath; | 2027 core.String gcsPath; |
1962 /** Required. The job name to use for the created job. */ | 2028 /** Required. The job name to use for the created job. */ |
1963 core.String jobName; | 2029 core.String jobName; |
1964 /** The runtime parameters to pass to the job. */ | 2030 /** The runtime parameters to pass to the job. */ |
1965 core.Map<core.String, core.String> parameters; | 2031 core.Map<core.String, core.String> parameters; |
1966 | 2032 |
1967 CreateJobFromTemplateRequest(); | 2033 CreateJobFromTemplateRequest(); |
1968 | 2034 |
1969 CreateJobFromTemplateRequest.fromJson(core.Map _json) { | 2035 CreateJobFromTemplateRequest.fromJson(core.Map _json) { |
(...skipping 49 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2019 } | 2085 } |
2020 return _json; | 2086 return _json; |
2021 } | 2087 } |
2022 } | 2088 } |
2023 | 2089 |
2024 /** Data disk assignment for a given VM instance. */ | 2090 /** Data disk assignment for a given VM instance. */ |
2025 class DataDiskAssignment { | 2091 class DataDiskAssignment { |
2026 /** | 2092 /** |
2027 * Mounted data disks. The order is important a data disk's 0-based index in | 2093 * Mounted data disks. The order is important a data disk's 0-based index in |
2028 * this list defines which persistent directory the disk is mounted to, for | 2094 * this list defines which persistent directory the disk is mounted to, for |
2029 * example the list of { "myproject-1014-104817-4c2-harness-0-disk-0" }, { | 2095 * example the list of { "myproject-1014-104817-4c2-harness-0-disk-0" }, |
2030 * "myproject-1014-104817-4c2-harness-0-disk-1" }. | 2096 * { "myproject-1014-104817-4c2-harness-0-disk-1" }. |
2031 */ | 2097 */ |
2032 core.List<core.String> dataDisks; | 2098 core.List<core.String> dataDisks; |
2033 /** | 2099 /** |
2034 * VM instance name the data disks mounted to, for example | 2100 * VM instance name the data disks mounted to, for example |
2035 * "myproject-1014-104817-4c2-harness-0". | 2101 * "myproject-1014-104817-4c2-harness-0". |
2036 */ | 2102 */ |
2037 core.String vmInstance; | 2103 core.String vmInstance; |
2038 | 2104 |
2039 DataDiskAssignment(); | 2105 DataDiskAssignment(); |
2040 | 2106 |
(...skipping 12 matching lines...) Expand all Loading... |
2053 _json["dataDisks"] = dataDisks; | 2119 _json["dataDisks"] = dataDisks; |
2054 } | 2120 } |
2055 if (vmInstance != null) { | 2121 if (vmInstance != null) { |
2056 _json["vmInstance"] = vmInstance; | 2122 _json["vmInstance"] = vmInstance; |
2057 } | 2123 } |
2058 return _json; | 2124 return _json; |
2059 } | 2125 } |
2060 } | 2126 } |
2061 | 2127 |
2062 /** | 2128 /** |
2063 * Specification of one of the bundles produced as a result of splitting a | 2129 * Specification of one of the bundles produced as a result of splitting |
2064 * Source (e.g. when executing a SourceSplitRequest, or when splitting an active | 2130 * a Source (e.g. when executing a SourceSplitRequest, or when |
2065 * task using WorkItemStatus.dynamic_source_split), relative to the source being | 2131 * splitting an active task using WorkItemStatus.dynamic_source_split), |
2066 * split. | 2132 * relative to the source being split. |
2067 */ | 2133 */ |
2068 class DerivedSource { | 2134 class DerivedSource { |
2069 /** | 2135 /** |
2070 * What source to base the produced source on (if any). | 2136 * What source to base the produced source on (if any). |
2071 * Possible string values are: | 2137 * Possible string values are: |
2072 * - "SOURCE_DERIVATION_MODE_UNKNOWN" : A SOURCE_DERIVATION_MODE_UNKNOWN. | 2138 * - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or |
2073 * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : A | 2139 * unspecified. |
2074 * SOURCE_DERIVATION_MODE_INDEPENDENT. | 2140 * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent |
2075 * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : A | 2141 * Source with no base. |
2076 * SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT. | 2142 * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on the |
2077 * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : A | 2143 * Source being split. |
2078 * SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT. | 2144 * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on |
| 2145 * the base of the Source being split. |
2079 */ | 2146 */ |
2080 core.String derivationMode; | 2147 core.String derivationMode; |
2081 /** Specification of the source. */ | 2148 /** Specification of the source. */ |
2082 Source source; | 2149 Source source; |
2083 | 2150 |
2084 DerivedSource(); | 2151 DerivedSource(); |
2085 | 2152 |
2086 DerivedSource.fromJson(core.Map _json) { | 2153 DerivedSource.fromJson(core.Map _json) { |
2087 if (_json.containsKey("derivationMode")) { | 2154 if (_json.containsKey("derivationMode")) { |
2088 derivationMode = _json["derivationMode"]; | 2155 derivationMode = _json["derivationMode"]; |
(...skipping 11 matching lines...) Expand all Loading... |
2100 if (source != null) { | 2167 if (source != null) { |
2101 _json["source"] = (source).toJson(); | 2168 _json["source"] = (source).toJson(); |
2102 } | 2169 } |
2103 return _json; | 2170 return _json; |
2104 } | 2171 } |
2105 } | 2172 } |
2106 | 2173 |
2107 /** Describes the data disk used by a workflow job. */ | 2174 /** Describes the data disk used by a workflow job. */ |
2108 class Disk { | 2175 class Disk { |
2109 /** | 2176 /** |
2110 * Disk storage type, as defined by Google Compute Engine. This must be a disk | 2177 * Disk storage type, as defined by Google Compute Engine. This |
2111 * type appropriate to the project and zone in which the workers will run. If | 2178 * must be a disk type appropriate to the project and zone in which |
2112 * unknown or unspecified, the service will attempt to choose a reasonable | 2179 * the workers will run. If unknown or unspecified, the service |
2113 * default. For example, the standard persistent disk type is a resource name | 2180 * will attempt to choose a reasonable default. |
2114 * typically ending in "pd-standard". If SSD persistent disks are available, | 2181 * |
2115 * the resource name typically ends with "pd-ssd". The actual valid values are | 2182 * For example, the standard persistent disk type is a resource name |
2116 * defined the Google Compute Engine API, not by the Cloud Dataflow API; | 2183 * typically ending in "pd-standard". If SSD persistent disks are |
2117 * consult the Google Compute Engine documentation for more information about | 2184 * available, the resource name typically ends with "pd-ssd". The |
2118 * determining the set of available disk types for a particular project and | 2185 * actual valid values are defined the Google Compute Engine API, |
2119 * zone. Google Compute Engine Disk types are local to a particular project in | 2186 * not by the Cloud Dataflow API; consult the Google Compute Engine |
2120 * a particular zone, and so the resource name will typically look something | 2187 * documentation for more information about determining the set of |
2121 * like this: | 2188 * available disk types for a particular project and zone. |
| 2189 * |
| 2190 * Google Compute Engine Disk types are local to a particular |
| 2191 * project in a particular zone, and so the resource name will |
| 2192 * typically look something like this: |
| 2193 * |
2122 * compute.googleapis.com/projects/project-id/zones/zone/diskTypes/pd-standard | 2194 * compute.googleapis.com/projects/project-id/zones/zone/diskTypes/pd-standard |
2123 */ | 2195 */ |
2124 core.String diskType; | 2196 core.String diskType; |
2125 /** Directory in a VM where disk is mounted. */ | 2197 /** Directory in a VM where disk is mounted. */ |
2126 core.String mountPoint; | 2198 core.String mountPoint; |
2127 /** | 2199 /** |
2128 * Size of disk in GB. If zero or unspecified, the service will attempt to | 2200 * Size of disk in GB. If zero or unspecified, the service will |
2129 * choose a reasonable default. | 2201 * attempt to choose a reasonable default. |
2130 */ | 2202 */ |
2131 core.int sizeGb; | 2203 core.int sizeGb; |
2132 | 2204 |
2133 Disk(); | 2205 Disk(); |
2134 | 2206 |
2135 Disk.fromJson(core.Map _json) { | 2207 Disk.fromJson(core.Map _json) { |
2136 if (_json.containsKey("diskType")) { | 2208 if (_json.containsKey("diskType")) { |
2137 diskType = _json["diskType"]; | 2209 diskType = _json["diskType"]; |
2138 } | 2210 } |
2139 if (_json.containsKey("mountPoint")) { | 2211 if (_json.containsKey("mountPoint")) { |
(...skipping 70 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2210 _json["sum"] = (sum).toJson(); | 2282 _json["sum"] = (sum).toJson(); |
2211 } | 2283 } |
2212 if (sumOfSquares != null) { | 2284 if (sumOfSquares != null) { |
2213 _json["sumOfSquares"] = sumOfSquares; | 2285 _json["sumOfSquares"] = sumOfSquares; |
2214 } | 2286 } |
2215 return _json; | 2287 return _json; |
2216 } | 2288 } |
2217 } | 2289 } |
2218 | 2290 |
2219 /** | 2291 /** |
2220 * When a task splits using WorkItemStatus.dynamic_source_split, this message | 2292 * When a task splits using WorkItemStatus.dynamic_source_split, this |
2221 * describes the two parts of the split relative to the description of the | 2293 * message describes the two parts of the split relative to the |
2222 * current task's input. | 2294 * description of the current task's input. |
2223 */ | 2295 */ |
2224 class DynamicSourceSplit { | 2296 class DynamicSourceSplit { |
2225 /** | 2297 /** |
2226 * Primary part (continued to be processed by worker). Specified relative to | 2298 * Primary part (continued to be processed by worker). |
2227 * the previously-current source. Becomes current. | 2299 * Specified relative to the previously-current source. |
| 2300 * Becomes current. |
2228 */ | 2301 */ |
2229 DerivedSource primary; | 2302 DerivedSource primary; |
2230 /** | 2303 /** |
2231 * Residual part (returned to the pool of work). Specified relative to the | 2304 * Residual part (returned to the pool of work). |
2232 * previously-current source. | 2305 * Specified relative to the previously-current source. |
2233 */ | 2306 */ |
2234 DerivedSource residual; | 2307 DerivedSource residual; |
2235 | 2308 |
2236 DynamicSourceSplit(); | 2309 DynamicSourceSplit(); |
2237 | 2310 |
2238 DynamicSourceSplit.fromJson(core.Map _json) { | 2311 DynamicSourceSplit.fromJson(core.Map _json) { |
2239 if (_json.containsKey("primary")) { | 2312 if (_json.containsKey("primary")) { |
2240 primary = new DerivedSource.fromJson(_json["primary"]); | 2313 primary = new DerivedSource.fromJson(_json["primary"]); |
2241 } | 2314 } |
2242 if (_json.containsKey("residual")) { | 2315 if (_json.containsKey("residual")) { |
2243 residual = new DerivedSource.fromJson(_json["residual"]); | 2316 residual = new DerivedSource.fromJson(_json["residual"]); |
2244 } | 2317 } |
2245 } | 2318 } |
2246 | 2319 |
2247 core.Map toJson() { | 2320 core.Map toJson() { |
2248 var _json = new core.Map(); | 2321 var _json = new core.Map(); |
2249 if (primary != null) { | 2322 if (primary != null) { |
2250 _json["primary"] = (primary).toJson(); | 2323 _json["primary"] = (primary).toJson(); |
2251 } | 2324 } |
2252 if (residual != null) { | 2325 if (residual != null) { |
2253 _json["residual"] = (residual).toJson(); | 2326 _json["residual"] = (residual).toJson(); |
2254 } | 2327 } |
2255 return _json; | 2328 return _json; |
2256 } | 2329 } |
2257 } | 2330 } |
2258 | 2331 |
2259 /** Describes the environment in which a Dataflow Job runs. */ | 2332 /** Describes the environment in which a Dataflow Job runs. */ |
2260 class Environment { | 2333 class Environment { |
2261 /** | 2334 /** |
2262 * The type of cluster manager API to use. If unknown or unspecified, the | 2335 * The type of cluster manager API to use. If unknown or |
2263 * service will attempt to choose a reasonable default. This should be in the | 2336 * unspecified, the service will attempt to choose a reasonable |
2264 * form of the API service name, e.g. "compute.googleapis.com". | 2337 * default. This should be in the form of the API service name, |
| 2338 * e.g. "compute.googleapis.com". |
2265 */ | 2339 */ |
2266 core.String clusterManagerApiService; | 2340 core.String clusterManagerApiService; |
2267 /** | 2341 /** |
2268 * The dataset for the current project where various workflow related tables | 2342 * The dataset for the current project where various workflow |
2269 * are stored. The supported resource type is: Google BigQuery: | 2343 * related tables are stored. |
2270 * bigquery.googleapis.com/{dataset} | 2344 * |
| 2345 * The supported resource type is: |
| 2346 * |
| 2347 * Google BigQuery: |
| 2348 * bigquery.googleapis.com/{dataset} |
2271 */ | 2349 */ |
2272 core.String dataset; | 2350 core.String dataset; |
2273 /** The list of experiments to enable. */ | 2351 /** The list of experiments to enable. */ |
2274 core.List<core.String> experiments; | 2352 core.List<core.String> experiments; |
2275 /** | 2353 /** |
2276 * Experimental settings. | 2354 * Experimental settings. |
2277 * | 2355 * |
2278 * The values for Object must be JSON objects. It can consist of `num`, | 2356 * The values for Object must be JSON objects. It can consist of `num`, |
2279 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2357 * `String`, `bool` and `null` as well as `Map` and `List` values. |
2280 */ | 2358 */ |
2281 core.Map<core.String, core.Object> internalExperiments; | 2359 core.Map<core.String, core.Object> internalExperiments; |
2282 /** | 2360 /** |
2283 * The Cloud Dataflow SDK pipeline options specified by the user. These | 2361 * The Cloud Dataflow SDK pipeline options specified by the user. These |
2284 * options are passed through the service and are used to recreate the SDK | 2362 * options are passed through the service and are used to recreate the |
2285 * pipeline options on the worker in a language agnostic and platform | 2363 * SDK pipeline options on the worker in a language agnostic and platform |
2286 * independent way. | 2364 * independent way. |
2287 * | 2365 * |
2288 * The values for Object must be JSON objects. It can consist of `num`, | 2366 * The values for Object must be JSON objects. It can consist of `num`, |
2289 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2367 * `String`, `bool` and `null` as well as `Map` and `List` values. |
2290 */ | 2368 */ |
2291 core.Map<core.String, core.Object> sdkPipelineOptions; | 2369 core.Map<core.String, core.Object> sdkPipelineOptions; |
2292 /** Identity to run virtual machines as. Defaults to the default account. */ | 2370 /** Identity to run virtual machines as. Defaults to the default account. */ |
2293 core.String serviceAccountEmail; | 2371 core.String serviceAccountEmail; |
2294 /** | 2372 /** |
2295 * The prefix of the resources the system should use for temporary storage. | 2373 * The prefix of the resources the system should use for temporary |
2296 * The system will append the suffix "/temp-{JOBNAME} to this resource prefix, | 2374 * storage. The system will append the suffix "/temp-{JOBNAME} to |
2297 * where {JOBNAME} is the value of the job_name field. The resulting bucket | 2375 * this resource prefix, where {JOBNAME} is the value of the |
2298 * and object prefix is used as the prefix of the resources used to store | 2376 * job_name field. The resulting bucket and object prefix is used |
2299 * temporary data needed during the job execution. NOTE: This will override | 2377 * as the prefix of the resources used to store temporary data |
2300 * the value in taskrunner_settings. The supported resource type is: Google | 2378 * needed during the job execution. NOTE: This will override the |
2301 * Cloud Storage: storage.googleapis.com/{bucket}/{object} | 2379 * value in taskrunner_settings. |
2302 * bucket.storage.googleapis.com/{object} | 2380 * The supported resource type is: |
| 2381 * |
| 2382 * Google Cloud Storage: |
| 2383 * |
| 2384 * storage.googleapis.com/{bucket}/{object} |
| 2385 * bucket.storage.googleapis.com/{object} |
2303 */ | 2386 */ |
2304 core.String tempStoragePrefix; | 2387 core.String tempStoragePrefix; |
2305 /** | 2388 /** |
2306 * A description of the process that generated the request. | 2389 * A description of the process that generated the request. |
2307 * | 2390 * |
2308 * The values for Object must be JSON objects. It can consist of `num`, | 2391 * The values for Object must be JSON objects. It can consist of `num`, |
2309 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2392 * `String`, `bool` and `null` as well as `Map` and `List` values. |
2310 */ | 2393 */ |
2311 core.Map<core.String, core.Object> userAgent; | 2394 core.Map<core.String, core.Object> userAgent; |
2312 /** | 2395 /** |
2313 * A structure describing which components and their versions of the service | 2396 * A structure describing which components and their versions of the service |
2314 * are required in order to run the job. | 2397 * are required in order to run the job. |
2315 * | 2398 * |
2316 * The values for Object must be JSON objects. It can consist of `num`, | 2399 * The values for Object must be JSON objects. It can consist of `num`, |
2317 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2400 * `String`, `bool` and `null` as well as `Map` and `List` values. |
2318 */ | 2401 */ |
2319 core.Map<core.String, core.Object> version; | 2402 core.Map<core.String, core.Object> version; |
2320 /** | 2403 /** |
2321 * The worker pools. At least one "harness" worker pool must be specified in | 2404 * The worker pools. At least one "harness" worker pool must be |
2322 * order for the job to have workers. | 2405 * specified in order for the job to have workers. |
2323 */ | 2406 */ |
2324 core.List<WorkerPool> workerPools; | 2407 core.List<WorkerPool> workerPools; |
2325 | 2408 |
2326 Environment(); | 2409 Environment(); |
2327 | 2410 |
2328 Environment.fromJson(core.Map _json) { | 2411 Environment.fromJson(core.Map _json) { |
2329 if (_json.containsKey("clusterManagerApiService")) { | 2412 if (_json.containsKey("clusterManagerApiService")) { |
2330 clusterManagerApiService = _json["clusterManagerApiService"]; | 2413 clusterManagerApiService = _json["clusterManagerApiService"]; |
2331 } | 2414 } |
2332 if (_json.containsKey("dataset")) { | 2415 if (_json.containsKey("dataset")) { |
(...skipping 154 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2487 } | 2570 } |
2488 if (sum != null) { | 2571 if (sum != null) { |
2489 _json["sum"] = sum; | 2572 _json["sum"] = sum; |
2490 } | 2573 } |
2491 return _json; | 2574 return _json; |
2492 } | 2575 } |
2493 } | 2576 } |
2494 | 2577 |
2495 /** Request to get updated debug configuration for component. */ | 2578 /** Request to get updated debug configuration for component. */ |
2496 class GetDebugConfigRequest { | 2579 class GetDebugConfigRequest { |
2497 /** The internal component id for which debug configuration is requested. */ | 2580 /** |
| 2581 * The internal component id for which debug configuration is |
| 2582 * requested. |
| 2583 */ |
2498 core.String componentId; | 2584 core.String componentId; |
2499 /** The worker id, i.e., VM hostname. */ | 2585 /** The worker id, i.e., VM hostname. */ |
2500 core.String workerId; | 2586 core.String workerId; |
2501 | 2587 |
2502 GetDebugConfigRequest(); | 2588 GetDebugConfigRequest(); |
2503 | 2589 |
2504 GetDebugConfigRequest.fromJson(core.Map _json) { | 2590 GetDebugConfigRequest.fromJson(core.Map _json) { |
2505 if (_json.containsKey("componentId")) { | 2591 if (_json.containsKey("componentId")) { |
2506 componentId = _json["componentId"]; | 2592 componentId = _json["componentId"]; |
2507 } | 2593 } |
(...skipping 30 matching lines...) Expand all Loading... |
2538 core.Map toJson() { | 2624 core.Map toJson() { |
2539 var _json = new core.Map(); | 2625 var _json = new core.Map(); |
2540 if (config != null) { | 2626 if (config != null) { |
2541 _json["config"] = config; | 2627 _json["config"] = config; |
2542 } | 2628 } |
2543 return _json; | 2629 return _json; |
2544 } | 2630 } |
2545 } | 2631 } |
2546 | 2632 |
2547 /** | 2633 /** |
2548 * An input of an instruction, as a reference to an output of a producer | 2634 * An input of an instruction, as a reference to an output of a |
2549 * instruction. | 2635 * producer instruction. |
2550 */ | 2636 */ |
2551 class InstructionInput { | 2637 class InstructionInput { |
2552 /** The output index (origin zero) within the producer. */ | 2638 /** The output index (origin zero) within the producer. */ |
2553 core.int outputNum; | 2639 core.int outputNum; |
2554 /** | 2640 /** |
2555 * The index (origin zero) of the parallel instruction that produces the | 2641 * The index (origin zero) of the parallel instruction that produces |
2556 * output to be consumed by this input. This index is relative to the list of | 2642 * the output to be consumed by this input. This index is relative |
2557 * instructions in this input's instruction's containing MapTask. | 2643 * to the list of instructions in this input's instruction's |
| 2644 * containing MapTask. |
2558 */ | 2645 */ |
2559 core.int producerInstructionIndex; | 2646 core.int producerInstructionIndex; |
2560 | 2647 |
2561 InstructionInput(); | 2648 InstructionInput(); |
2562 | 2649 |
2563 InstructionInput.fromJson(core.Map _json) { | 2650 InstructionInput.fromJson(core.Map _json) { |
2564 if (_json.containsKey("outputNum")) { | 2651 if (_json.containsKey("outputNum")) { |
2565 outputNum = _json["outputNum"]; | 2652 outputNum = _json["outputNum"]; |
2566 } | 2653 } |
2567 if (_json.containsKey("producerInstructionIndex")) { | 2654 if (_json.containsKey("producerInstructionIndex")) { |
(...skipping 28 matching lines...) Expand all Loading... |
2596 * For system-generated byte and mean byte metrics, certain instructions | 2683 * For system-generated byte and mean byte metrics, certain instructions |
2597 * should only report the key size. | 2684 * should only report the key size. |
2598 */ | 2685 */ |
2599 core.bool onlyCountKeyBytes; | 2686 core.bool onlyCountKeyBytes; |
2600 /** | 2687 /** |
2601 * For system-generated byte and mean byte metrics, certain instructions | 2688 * For system-generated byte and mean byte metrics, certain instructions |
2602 * should only report the value size. | 2689 * should only report the value size. |
2603 */ | 2690 */ |
2604 core.bool onlyCountValueBytes; | 2691 core.bool onlyCountValueBytes; |
2605 /** | 2692 /** |
2606 * System-defined name for this output in the original workflow graph. Outputs | 2693 * System-defined name for this output in the original workflow graph. |
2607 * that do not contribute to an original instruction do not set this. | 2694 * Outputs that do not contribute to an original instruction do not set this. |
2608 */ | 2695 */ |
2609 core.String originalName; | 2696 core.String originalName; |
2610 /** System-defined name of this output. Unique across the workflow. */ | 2697 /** |
| 2698 * System-defined name of this output. |
| 2699 * Unique across the workflow. |
| 2700 */ |
2611 core.String systemName; | 2701 core.String systemName; |
2612 | 2702 |
2613 InstructionOutput(); | 2703 InstructionOutput(); |
2614 | 2704 |
2615 InstructionOutput.fromJson(core.Map _json) { | 2705 InstructionOutput.fromJson(core.Map _json) { |
2616 if (_json.containsKey("codec")) { | 2706 if (_json.containsKey("codec")) { |
2617 codec = _json["codec"]; | 2707 codec = _json["codec"]; |
2618 } | 2708 } |
2619 if (_json.containsKey("name")) { | 2709 if (_json.containsKey("name")) { |
2620 name = _json["name"]; | 2710 name = _json["name"]; |
(...skipping 85 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2706 _json["sum"] = (sum).toJson(); | 2796 _json["sum"] = (sum).toJson(); |
2707 } | 2797 } |
2708 return _json; | 2798 return _json; |
2709 } | 2799 } |
2710 } | 2800 } |
2711 | 2801 |
2712 /** Defines a job to be run by the Cloud Dataflow service. */ | 2802 /** Defines a job to be run by the Cloud Dataflow service. */ |
2713 class Job { | 2803 class Job { |
2714 /** | 2804 /** |
2715 * The client's unique identifier of the job, re-used across retried attempts. | 2805 * The client's unique identifier of the job, re-used across retried attempts. |
2716 * If this field is set, the service will ensure its uniqueness. The request | 2806 * If this field is set, the service will ensure its uniqueness. |
2717 * to create a job will fail if the service has knowledge of a previously | 2807 * The request to create a job will fail if the service has knowledge of a |
2718 * submitted job with the same client's ID and job name. The caller may use | 2808 * previously submitted job with the same client's ID and job name. |
2719 * this field to ensure idempotence of job creation across retried attempts to | 2809 * The caller may use this field to ensure idempotence of job |
2720 * create a job. By default, the field is empty and, in that case, the service | 2810 * creation across retried attempts to create a job. |
2721 * ignores it. | 2811 * By default, the field is empty and, in that case, the service ignores it. |
2722 */ | 2812 */ |
2723 core.String clientRequestId; | 2813 core.String clientRequestId; |
2724 /** | 2814 /** |
2725 * The timestamp when the job was initially created. Immutable and set by the | 2815 * The timestamp when the job was initially created. Immutable and set by the |
2726 * Cloud Dataflow service. | 2816 * Cloud Dataflow service. |
2727 */ | 2817 */ |
2728 core.String createTime; | 2818 core.String createTime; |
2729 /** | 2819 /** |
2730 * The current state of the job. Jobs are created in the `JOB_STATE_STOPPED` | 2820 * The current state of the job. |
2731 * state unless otherwise specified. A job in the `JOB_STATE_RUNNING` state | 2821 * |
2732 * may asynchronously enter a terminal state. After a job has reached a | 2822 * Jobs are created in the `JOB_STATE_STOPPED` state unless otherwise |
2733 * terminal state, no further state updates may be made. This field may be | 2823 * specified. |
2734 * mutated by the Cloud Dataflow service; callers cannot mutate it. | 2824 * |
| 2825 * A job in the `JOB_STATE_RUNNING` state may asynchronously enter a |
| 2826 * terminal state. After a job has reached a terminal state, no |
| 2827 * further state updates may be made. |
| 2828 * |
| 2829 * This field may be mutated by the Cloud Dataflow service; |
| 2830 * callers cannot mutate it. |
2735 * Possible string values are: | 2831 * Possible string values are: |
2736 * - "JOB_STATE_UNKNOWN" : A JOB_STATE_UNKNOWN. | 2832 * - "JOB_STATE_UNKNOWN" : The job's run state isn't specified. |
2737 * - "JOB_STATE_STOPPED" : A JOB_STATE_STOPPED. | 2833 * - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has not |
2738 * - "JOB_STATE_RUNNING" : A JOB_STATE_RUNNING. | 2834 * yet started to run. |
2739 * - "JOB_STATE_DONE" : A JOB_STATE_DONE. | 2835 * - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is |
2740 * - "JOB_STATE_FAILED" : A JOB_STATE_FAILED. | 2836 * currently running. |
2741 * - "JOB_STATE_CANCELLED" : A JOB_STATE_CANCELLED. | 2837 * - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has |
2742 * - "JOB_STATE_UPDATED" : A JOB_STATE_UPDATED. | 2838 * successfully completed. |
2743 * - "JOB_STATE_DRAINING" : A JOB_STATE_DRAINING. | 2839 * This is a terminal job state. This state may be set by the Cloud Dataflow |
2744 * - "JOB_STATE_DRAINED" : A JOB_STATE_DRAINED. | 2840 * service, as a transition from `JOB_STATE_RUNNING`. It may also be set via a |
| 2841 * Cloud Dataflow `UpdateJob` call, if the job has not yet reached a terminal |
| 2842 * state. |
| 2843 * - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has |
| 2844 * failed. This is a |
| 2845 * terminal job state. This state may only be set by the Cloud Dataflow |
| 2846 * service, and only as a transition from `JOB_STATE_RUNNING`. |
| 2847 * - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job has |
| 2848 * been explicitly |
| 2849 * cancelled. This is a terminal job state. This state may only be |
| 2850 * set via a Cloud Dataflow `UpdateJob` call, and only if the job has not |
| 2851 * yet reached another terminal state. |
| 2852 * - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was |
| 2853 * successfully updated, |
| 2854 * meaning that this job was stopped and another job was started, inheriting |
| 2855 * state from this one. This is a terminal job state. This state may only be |
| 2856 * set by the Cloud Dataflow service, and only as a transition from |
| 2857 * `JOB_STATE_RUNNING`. |
| 2858 * - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is in |
| 2859 * the process of draining. |
| 2860 * A draining job has stopped pulling from its input sources and is processing |
| 2861 * any data that remains in-flight. This state may be set via a Cloud Dataflow |
| 2862 * `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs |
| 2863 * that are draining may only transition to `JOB_STATE_DRAINED`, |
| 2864 * `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`. |
| 2865 * - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has been |
| 2866 * drained. |
| 2867 * A drained job terminated by stopping pulling from its input sources and |
| 2868 * processing any data that remained in-flight when draining was requested. |
| 2869 * This state is a terminal state, may only be set by the Cloud Dataflow |
| 2870 * service, and only as a transition from `JOB_STATE_DRAINING`. |
2745 */ | 2871 */ |
2746 core.String currentState; | 2872 core.String currentState; |
2747 /** The timestamp associated with the current state. */ | 2873 /** The timestamp associated with the current state. */ |
2748 core.String currentStateTime; | 2874 core.String currentStateTime; |
2749 /** The environment for the job. */ | 2875 /** The environment for the job. */ |
2750 Environment environment; | 2876 Environment environment; |
2751 /** Information about how the Cloud Dataflow service will run the job. */ | 2877 /** Information about how the Cloud Dataflow service will run the job. */ |
2752 JobExecutionInfo executionInfo; | 2878 JobExecutionInfo executionInfo; |
2753 /** | 2879 /** |
2754 * The unique ID of this job. This field is set by the Cloud Dataflow service | 2880 * The unique ID of this job. |
2755 * when the Job is created, and is immutable for the life of the job. | 2881 * |
| 2882 * This field is set by the Cloud Dataflow service when the Job is |
| 2883 * created, and is immutable for the life of the job. |
2756 */ | 2884 */ |
2757 core.String id; | 2885 core.String id; |
2758 /** | 2886 /** |
2759 * User-defined labels for this job. The labels map can contain no more than | 2887 * User-defined labels for this job. |
2760 * 64 entries. Entries of the labels map are UTF8 strings that comply with the | 2888 * |
2761 * following restrictions: * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} * | 2889 * The labels map can contain no more than 64 entries. Entries of the labels |
2762 * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} * Both keys and | 2890 * map are UTF8 strings that comply with the following restrictions: |
2763 * values are additionally constrained to be <= 128 bytes in size. | 2891 * |
| 2892 * * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62} |
| 2893 * * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63} |
| 2894 * * Both keys and values are additionally constrained to be <= 128 bytes in |
| 2895 * size. |
2764 */ | 2896 */ |
2765 core.Map<core.String, core.String> labels; | 2897 core.Map<core.String, core.String> labels; |
2766 /** The location that contains this job. */ | 2898 /** The location that contains this job. */ |
2767 core.String location; | 2899 core.String location; |
2768 /** | 2900 /** |
2769 * The user-specified Cloud Dataflow job name. Only one Job with a given name | 2901 * The user-specified Cloud Dataflow job name. |
2770 * may exist in a project at any given time. If a caller attempts to create a | 2902 * |
2771 * Job with the same name as an already-existing Job, the attempt returns the | 2903 * Only one Job with a given name may exist in a project at any |
2772 * existing Job. The name must match the regular expression | 2904 * given time. If a caller attempts to create a Job with the same |
| 2905 * name as an already-existing Job, the attempt returns the |
| 2906 * existing Job. |
| 2907 * |
| 2908 * The name must match the regular expression |
2773 * `[a-z]([-a-z0-9]{0,38}[a-z0-9])?` | 2909 * `[a-z]([-a-z0-9]{0,38}[a-z0-9])?` |
2774 */ | 2910 */ |
2775 core.String name; | 2911 core.String name; |
2776 /** The ID of the Cloud Platform project that the job belongs to. */ | 2912 /** The ID of the Cloud Platform project that the job belongs to. */ |
2777 core.String projectId; | 2913 core.String projectId; |
2778 /** | 2914 /** |
2779 * If this job is an update of an existing job, this field is the job ID of | 2915 * If this job is an update of an existing job, this field is the job ID |
2780 * the job it replaced. When sending a `CreateJobRequest`, you can update a | 2916 * of the job it replaced. |
2781 * job by specifying it here. The job named here is stopped, and its | 2917 * |
2782 * intermediate state is transferred to this job. | 2918 * When sending a `CreateJobRequest`, you can update a job by specifying it |
| 2919 * here. The job named here is stopped, and its intermediate state is |
| 2920 * transferred to this job. |
2783 */ | 2921 */ |
2784 core.String replaceJobId; | 2922 core.String replaceJobId; |
2785 /** | 2923 /** |
2786 * If another job is an update of this job (and thus, this job is in | 2924 * If another job is an update of this job (and thus, this job is in |
2787 * `JOB_STATE_UPDATED`), this field contains the ID of that job. | 2925 * `JOB_STATE_UPDATED`), this field contains the ID of that job. |
2788 */ | 2926 */ |
2789 core.String replacedByJobId; | 2927 core.String replacedByJobId; |
2790 /** | 2928 /** |
2791 * The job's requested state. `UpdateJob` may be used to switch between the | 2929 * The job's requested state. |
2792 * `JOB_STATE_STOPPED` and `JOB_STATE_RUNNING` states, by setting | 2930 * |
2793 * requested_state. `UpdateJob` may also be used to directly set a job's | 2931 * `UpdateJob` may be used to switch between the `JOB_STATE_STOPPED` and |
2794 * requested state to `JOB_STATE_CANCELLED` or `JOB_STATE_DONE`, irrevocably | 2932 * `JOB_STATE_RUNNING` states, by setting requested_state. `UpdateJob` may |
2795 * terminating the job if it has not already reached a terminal state. | 2933 * also be used to directly set a job's requested state to |
| 2934 * `JOB_STATE_CANCELLED` or `JOB_STATE_DONE`, irrevocably terminating the |
| 2935 * job if it has not already reached a terminal state. |
2796 * Possible string values are: | 2936 * Possible string values are: |
2797 * - "JOB_STATE_UNKNOWN" : A JOB_STATE_UNKNOWN. | 2937 * - "JOB_STATE_UNKNOWN" : The job's run state isn't specified. |
2798 * - "JOB_STATE_STOPPED" : A JOB_STATE_STOPPED. | 2938 * - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has not |
2799 * - "JOB_STATE_RUNNING" : A JOB_STATE_RUNNING. | 2939 * yet started to run. |
2800 * - "JOB_STATE_DONE" : A JOB_STATE_DONE. | 2940 * - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is |
2801 * - "JOB_STATE_FAILED" : A JOB_STATE_FAILED. | 2941 * currently running. |
2802 * - "JOB_STATE_CANCELLED" : A JOB_STATE_CANCELLED. | 2942 * - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has |
2803 * - "JOB_STATE_UPDATED" : A JOB_STATE_UPDATED. | 2943 * successfully completed. |
2804 * - "JOB_STATE_DRAINING" : A JOB_STATE_DRAINING. | 2944 * This is a terminal job state. This state may be set by the Cloud Dataflow |
2805 * - "JOB_STATE_DRAINED" : A JOB_STATE_DRAINED. | 2945 * service, as a transition from `JOB_STATE_RUNNING`. It may also be set via a |
| 2946 * Cloud Dataflow `UpdateJob` call, if the job has not yet reached a terminal |
| 2947 * state. |
| 2948 * - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has |
| 2949 * failed. This is a |
| 2950 * terminal job state. This state may only be set by the Cloud Dataflow |
| 2951 * service, and only as a transition from `JOB_STATE_RUNNING`. |
| 2952 * - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job has |
| 2953 * been explicitly |
| 2954 * cancelled. This is a terminal job state. This state may only be |
| 2955 * set via a Cloud Dataflow `UpdateJob` call, and only if the job has not |
| 2956 * yet reached another terminal state. |
| 2957 * - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was |
| 2958 * successfully updated, |
| 2959 * meaning that this job was stopped and another job was started, inheriting |
| 2960 * state from this one. This is a terminal job state. This state may only be |
| 2961 * set by the Cloud Dataflow service, and only as a transition from |
| 2962 * `JOB_STATE_RUNNING`. |
| 2963 * - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is in |
| 2964 * the process of draining. |
| 2965 * A draining job has stopped pulling from its input sources and is processing |
| 2966 * any data that remains in-flight. This state may be set via a Cloud Dataflow |
| 2967 * `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs |
| 2968 * that are draining may only transition to `JOB_STATE_DRAINED`, |
| 2969 * `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`. |
| 2970 * - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has been |
| 2971 * drained. |
| 2972 * A drained job terminated by stopping pulling from its input sources and |
| 2973 * processing any data that remained in-flight when draining was requested. |
| 2974 * This state is a terminal state, may only be set by the Cloud Dataflow |
| 2975 * service, and only as a transition from `JOB_STATE_DRAINING`. |
2806 */ | 2976 */ |
2807 core.String requestedState; | 2977 core.String requestedState; |
2808 /** The top-level steps that constitute the entire job. */ | 2978 /** The top-level steps that constitute the entire job. */ |
2809 core.List<Step> steps; | 2979 core.List<Step> steps; |
2810 /** | 2980 /** |
2811 * A set of files the system should be aware of that are used for temporary | 2981 * A set of files the system should be aware of that are used |
2812 * storage. These temporary files will be removed on job completion. No | 2982 * for temporary storage. These temporary files will be |
2813 * duplicates are allowed. No file patterns are supported. The supported files | 2983 * removed on job completion. |
2814 * are: Google Cloud Storage: storage.googleapis.com/{bucket}/{object} | 2984 * No duplicates are allowed. |
2815 * bucket.storage.googleapis.com/{object} | 2985 * No file patterns are supported. |
| 2986 * |
| 2987 * The supported files are: |
| 2988 * |
| 2989 * Google Cloud Storage: |
| 2990 * |
| 2991 * storage.googleapis.com/{bucket}/{object} |
| 2992 * bucket.storage.googleapis.com/{object} |
2816 */ | 2993 */ |
2817 core.List<core.String> tempFiles; | 2994 core.List<core.String> tempFiles; |
2818 /** | 2995 /** |
2819 * The map of transform name prefixes of the job to be replaced to the | 2996 * The map of transform name prefixes of the job to be replaced to the |
2820 * corresponding name prefixes of the new job. | 2997 * corresponding name prefixes of the new job. |
2821 */ | 2998 */ |
2822 core.Map<core.String, core.String> transformNameMapping; | 2999 core.Map<core.String, core.String> transformNameMapping; |
2823 /** | 3000 /** |
2824 * The type of Cloud Dataflow job. | 3001 * The type of Cloud Dataflow job. |
2825 * Possible string values are: | 3002 * Possible string values are: |
2826 * - "JOB_TYPE_UNKNOWN" : A JOB_TYPE_UNKNOWN. | 3003 * - "JOB_TYPE_UNKNOWN" : The type of the job is unspecified, or unknown. |
2827 * - "JOB_TYPE_BATCH" : A JOB_TYPE_BATCH. | 3004 * - "JOB_TYPE_BATCH" : A batch job with a well-defined end point: data is |
2828 * - "JOB_TYPE_STREAMING" : A JOB_TYPE_STREAMING. | 3005 * read, data is |
| 3006 * processed, data is written, and the job is done. |
| 3007 * - "JOB_TYPE_STREAMING" : A continuously streaming job with no end: data is |
| 3008 * read, |
| 3009 * processed, and written continuously. |
2829 */ | 3010 */ |
2830 core.String type; | 3011 core.String type; |
2831 | 3012 |
2832 Job(); | 3013 Job(); |
2833 | 3014 |
2834 Job.fromJson(core.Map _json) { | 3015 Job.fromJson(core.Map _json) { |
2835 if (_json.containsKey("clientRequestId")) { | 3016 if (_json.containsKey("clientRequestId")) { |
2836 clientRequestId = _json["clientRequestId"]; | 3017 clientRequestId = _json["clientRequestId"]; |
2837 } | 3018 } |
2838 if (_json.containsKey("createTime")) { | 3019 if (_json.containsKey("createTime")) { |
(...skipping 128 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2967 core.Map toJson() { | 3148 core.Map toJson() { |
2968 var _json = new core.Map(); | 3149 var _json = new core.Map(); |
2969 if (stages != null) { | 3150 if (stages != null) { |
2970 _json["stages"] = commons.mapMap(stages, (item) => (item).toJson()); | 3151 _json["stages"] = commons.mapMap(stages, (item) => (item).toJson()); |
2971 } | 3152 } |
2972 return _json; | 3153 return _json; |
2973 } | 3154 } |
2974 } | 3155 } |
2975 | 3156 |
2976 /** | 3157 /** |
2977 * Contains information about how a particular google.dataflow.v1beta3.Step will | 3158 * Contains information about how a particular |
2978 * be executed. | 3159 * google.dataflow.v1beta3.Step will be executed. |
2979 */ | 3160 */ |
2980 class JobExecutionStageInfo { | 3161 class JobExecutionStageInfo { |
2981 /** | 3162 /** |
2982 * The steps associated with the execution stage. Note that stages may have | 3163 * The steps associated with the execution stage. |
2983 * several steps, and that a given step might be run by more than one stage. | 3164 * Note that stages may have several steps, and that a given step |
| 3165 * might be run by more than one stage. |
2984 */ | 3166 */ |
2985 core.List<core.String> stepName; | 3167 core.List<core.String> stepName; |
2986 | 3168 |
2987 JobExecutionStageInfo(); | 3169 JobExecutionStageInfo(); |
2988 | 3170 |
2989 JobExecutionStageInfo.fromJson(core.Map _json) { | 3171 JobExecutionStageInfo.fromJson(core.Map _json) { |
2990 if (_json.containsKey("stepName")) { | 3172 if (_json.containsKey("stepName")) { |
2991 stepName = _json["stepName"]; | 3173 stepName = _json["stepName"]; |
2992 } | 3174 } |
2993 } | 3175 } |
2994 | 3176 |
2995 core.Map toJson() { | 3177 core.Map toJson() { |
2996 var _json = new core.Map(); | 3178 var _json = new core.Map(); |
2997 if (stepName != null) { | 3179 if (stepName != null) { |
2998 _json["stepName"] = stepName; | 3180 _json["stepName"] = stepName; |
2999 } | 3181 } |
3000 return _json; | 3182 return _json; |
3001 } | 3183 } |
3002 } | 3184 } |
3003 | 3185 |
3004 /** A particular message pertaining to a Dataflow job. */ | 3186 /** A particular message pertaining to a Dataflow job. */ |
3005 class JobMessage { | 3187 class JobMessage { |
3006 /** | 3188 /** |
3007 * Identifies the message. This is automatically generated by the service; the | 3189 * Identifies the message. This is automatically generated by the |
3008 * caller should treat it as an opaque string. | 3190 * service; the caller should treat it as an opaque string. |
3009 */ | 3191 */ |
3010 core.String id; | 3192 core.String id; |
3011 /** | 3193 /** |
3012 * Importance level of the message. | 3194 * Importance level of the message. |
3013 * Possible string values are: | 3195 * Possible string values are: |
3014 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN. | 3196 * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : The message importance isn't |
3015 * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG. | 3197 * specified, or is unknown. |
3016 * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED. | 3198 * - "JOB_MESSAGE_DEBUG" : The message is at the 'debug' level: typically only |
3017 * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC. | 3199 * useful for |
3018 * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING. | 3200 * software engineers working on the code the job is running. |
3019 * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR. | 3201 * Typically, Dataflow pipeline runners do not display log messages |
| 3202 * at this level by default. |
| 3203 * - "JOB_MESSAGE_DETAILED" : The message is at the 'detailed' level: somewhat |
| 3204 * verbose, but |
| 3205 * potentially useful to users. Typically, Dataflow pipeline |
| 3206 * runners do not display log messages at this level by default. |
| 3207 * These messages are displayed by default in the Dataflow |
| 3208 * monitoring UI. |
| 3209 * - "JOB_MESSAGE_BASIC" : The message is at the 'basic' level: useful for |
| 3210 * keeping |
| 3211 * track of the execution of a Dataflow pipeline. Typically, |
| 3212 * Dataflow pipeline runners display log messages at this level by |
| 3213 * default, and these messages are displayed by default in the |
| 3214 * Dataflow monitoring UI. |
| 3215 * - "JOB_MESSAGE_WARNING" : The message is at the 'warning' level: indicating |
| 3216 * a condition |
| 3217 * pertaining to a job which may require human intervention. |
| 3218 * Typically, Dataflow pipeline runners display log messages at this |
| 3219 * level by default, and these messages are displayed by default in |
| 3220 * the Dataflow monitoring UI. |
| 3221 * - "JOB_MESSAGE_ERROR" : The message is at the 'error' level: indicating a |
| 3222 * condition |
| 3223 * preventing a job from succeeding. Typically, Dataflow pipeline |
| 3224 * runners display log messages at this level by default, and these |
| 3225 * messages are displayed by default in the Dataflow monitoring UI. |
3020 */ | 3226 */ |
3021 core.String messageImportance; | 3227 core.String messageImportance; |
3022 /** The text of the message. */ | 3228 /** The text of the message. */ |
3023 core.String messageText; | 3229 core.String messageText; |
3024 /** The timestamp of the message. */ | 3230 /** The timestamp of the message. */ |
3025 core.String time; | 3231 core.String time; |
3026 | 3232 |
3027 JobMessage(); | 3233 JobMessage(); |
3028 | 3234 |
3029 JobMessage.fromJson(core.Map _json) { | 3235 JobMessage.fromJson(core.Map _json) { |
(...skipping 25 matching lines...) Expand all Loading... |
3055 if (time != null) { | 3261 if (time != null) { |
3056 _json["time"] = time; | 3262 _json["time"] = time; |
3057 } | 3263 } |
3058 return _json; | 3264 return _json; |
3059 } | 3265 } |
3060 } | 3266 } |
3061 | 3267 |
3062 /** | 3268 /** |
3063 * JobMetrics contains a collection of metrics descibing the detailed progress | 3269 * JobMetrics contains a collection of metrics descibing the detailed progress |
3064 * of a Dataflow job. Metrics correspond to user-defined and system-defined | 3270 * of a Dataflow job. Metrics correspond to user-defined and system-defined |
3065 * metrics in the job. This resource captures only the most recent values of | 3271 * metrics in the job. |
3066 * each metric; time-series data can be queried for them (under the same metric | 3272 * |
3067 * names) from Cloud Monitoring. | 3273 * This resource captures only the most recent values of each metric; |
| 3274 * time-series data can be queried for them (under the same metric names) |
| 3275 * from Cloud Monitoring. |
3068 */ | 3276 */ |
3069 class JobMetrics { | 3277 class JobMetrics { |
3070 /** Timestamp as of which metric values are current. */ | 3278 /** Timestamp as of which metric values are current. */ |
3071 core.String metricTime; | 3279 core.String metricTime; |
3072 /** All metrics for this job. */ | 3280 /** All metrics for this job. */ |
3073 core.List<MetricUpdate> metrics; | 3281 core.List<MetricUpdate> metrics; |
3074 | 3282 |
3075 JobMetrics(); | 3283 JobMetrics(); |
3076 | 3284 |
3077 JobMetrics.fromJson(core.Map _json) { | 3285 JobMetrics.fromJson(core.Map _json) { |
(...skipping 12 matching lines...) Expand all Loading... |
3090 } | 3298 } |
3091 if (metrics != null) { | 3299 if (metrics != null) { |
3092 _json["metrics"] = metrics.map((value) => (value).toJson()).toList(); | 3300 _json["metrics"] = metrics.map((value) => (value).toJson()).toList(); |
3093 } | 3301 } |
3094 return _json; | 3302 return _json; |
3095 } | 3303 } |
3096 } | 3304 } |
3097 | 3305 |
3098 /** | 3306 /** |
3099 * Data disk assignment information for a specific key-range of a sharded | 3307 * Data disk assignment information for a specific key-range of a sharded |
3100 * computation. Currently we only support UTF-8 character splits to simplify | 3308 * computation. |
3101 * encoding into JSON. | 3309 * Currently we only support UTF-8 character splits to simplify encoding into |
| 3310 * JSON. |
3102 */ | 3311 */ |
3103 class KeyRangeDataDiskAssignment { | 3312 class KeyRangeDataDiskAssignment { |
3104 /** | 3313 /** |
3105 * The name of the data disk where data for this range is stored. This name is | 3314 * The name of the data disk where data for this range is stored. |
3106 * local to the Google Cloud Platform project and uniquely identifies the disk | 3315 * This name is local to the Google Cloud Platform project and uniquely |
3107 * within that project, for example | 3316 * identifies the disk within that project, for example |
3108 * "myproject-1014-104817-4c2-harness-0-disk-1". | 3317 * "myproject-1014-104817-4c2-harness-0-disk-1". |
3109 */ | 3318 */ |
3110 core.String dataDisk; | 3319 core.String dataDisk; |
3111 /** The end (exclusive) of the key range. */ | 3320 /** The end (exclusive) of the key range. */ |
3112 core.String end; | 3321 core.String end; |
3113 /** The start (inclusive) of the key range. */ | 3322 /** The start (inclusive) of the key range. */ |
3114 core.String start; | 3323 core.String start; |
3115 | 3324 |
3116 KeyRangeDataDiskAssignment(); | 3325 KeyRangeDataDiskAssignment(); |
3117 | 3326 |
(...skipping 24 matching lines...) Expand all Loading... |
3142 } | 3351 } |
3143 } | 3352 } |
3144 | 3353 |
3145 /** | 3354 /** |
3146 * Location information for a specific key-range of a sharded computation. | 3355 * Location information for a specific key-range of a sharded computation. |
3147 * Currently we only support UTF-8 character splits to simplify encoding into | 3356 * Currently we only support UTF-8 character splits to simplify encoding into |
3148 * JSON. | 3357 * JSON. |
3149 */ | 3358 */ |
3150 class KeyRangeLocation { | 3359 class KeyRangeLocation { |
3151 /** | 3360 /** |
3152 * The name of the data disk where data for this range is stored. This name is | 3361 * The name of the data disk where data for this range is stored. |
3153 * local to the Google Cloud Platform project and uniquely identifies the disk | 3362 * This name is local to the Google Cloud Platform project and uniquely |
3154 * within that project, for example | 3363 * identifies the disk within that project, for example |
3155 * "myproject-1014-104817-4c2-harness-0-disk-1". | 3364 * "myproject-1014-104817-4c2-harness-0-disk-1". |
3156 */ | 3365 */ |
3157 core.String dataDisk; | 3366 core.String dataDisk; |
3158 /** | 3367 /** |
3159 * The physical location of this range assignment to be used for streaming | 3368 * The physical location of this range assignment to be used for |
3160 * computation cross-worker message delivery. | 3369 * streaming computation cross-worker message delivery. |
3161 */ | 3370 */ |
3162 core.String deliveryEndpoint; | 3371 core.String deliveryEndpoint; |
3163 /** The end (exclusive) of the key range. */ | 3372 /** The end (exclusive) of the key range. */ |
3164 core.String end; | 3373 core.String end; |
3165 /** | 3374 /** |
3166 * The location of the persistent state for this range, as a persistent | 3375 * The location of the persistent state for this range, as a |
3167 * directory in the worker local filesystem. | 3376 * persistent directory in the worker local filesystem. |
3168 */ | 3377 */ |
3169 core.String persistentDirectory; | 3378 core.String persistentDirectory; |
3170 /** The start (inclusive) of the key range. */ | 3379 /** The start (inclusive) of the key range. */ |
3171 core.String start; | 3380 core.String start; |
3172 | 3381 |
3173 KeyRangeLocation(); | 3382 KeyRangeLocation(); |
3174 | 3383 |
3175 KeyRangeLocation.fromJson(core.Map _json) { | 3384 KeyRangeLocation.fromJson(core.Map _json) { |
3176 if (_json.containsKey("dataDisk")) { | 3385 if (_json.containsKey("dataDisk")) { |
3177 dataDisk = _json["dataDisk"]; | 3386 dataDisk = _json["dataDisk"]; |
(...skipping 42 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3220 /** The initial lease period. */ | 3429 /** The initial lease period. */ |
3221 core.String requestedLeaseDuration; | 3430 core.String requestedLeaseDuration; |
3222 /** Filter for WorkItem type. */ | 3431 /** Filter for WorkItem type. */ |
3223 core.List<core.String> workItemTypes; | 3432 core.List<core.String> workItemTypes; |
3224 /** | 3433 /** |
3225 * Worker capabilities. WorkItems might be limited to workers with specific | 3434 * Worker capabilities. WorkItems might be limited to workers with specific |
3226 * capabilities. | 3435 * capabilities. |
3227 */ | 3436 */ |
3228 core.List<core.String> workerCapabilities; | 3437 core.List<core.String> workerCapabilities; |
3229 /** | 3438 /** |
3230 * Identifies the worker leasing work -- typically the ID of the virtual | 3439 * Identifies the worker leasing work -- typically the ID of the |
3231 * machine running the worker. | 3440 * virtual machine running the worker. |
3232 */ | 3441 */ |
3233 core.String workerId; | 3442 core.String workerId; |
3234 | 3443 |
3235 LeaseWorkItemRequest(); | 3444 LeaseWorkItemRequest(); |
3236 | 3445 |
3237 LeaseWorkItemRequest.fromJson(core.Map _json) { | 3446 LeaseWorkItemRequest.fromJson(core.Map _json) { |
3238 if (_json.containsKey("currentWorkerTime")) { | 3447 if (_json.containsKey("currentWorkerTime")) { |
3239 currentWorkerTime = _json["currentWorkerTime"]; | 3448 currentWorkerTime = _json["currentWorkerTime"]; |
3240 } | 3449 } |
3241 if (_json.containsKey("location")) { | 3450 if (_json.containsKey("location")) { |
(...skipping 83 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3325 _json["jobMessages"] = jobMessages.map((value) => (value).toJson()).toList
(); | 3534 _json["jobMessages"] = jobMessages.map((value) => (value).toJson()).toList
(); |
3326 } | 3535 } |
3327 if (nextPageToken != null) { | 3536 if (nextPageToken != null) { |
3328 _json["nextPageToken"] = nextPageToken; | 3537 _json["nextPageToken"] = nextPageToken; |
3329 } | 3538 } |
3330 return _json; | 3539 return _json; |
3331 } | 3540 } |
3332 } | 3541 } |
3333 | 3542 |
3334 /** | 3543 /** |
3335 * Response to a request to list Cloud Dataflow jobs. This may be a partial | 3544 * Response to a request to list Cloud Dataflow jobs. This may be a partial |
3336 * response, depending on the page size in the ListJobsRequest. | 3545 * response, depending on the page size in the ListJobsRequest. |
3337 */ | 3546 */ |
3338 class ListJobsResponse { | 3547 class ListJobsResponse { |
3339 /** Zero or more messages describing locations that failed to respond. */ | 3548 /** Zero or more messages describing locations that failed to respond. */ |
3340 core.List<FailedLocation> failedLocation; | 3549 core.List<FailedLocation> failedLocation; |
3341 /** A subset of the requested job information. */ | 3550 /** A subset of the requested job information. */ |
3342 core.List<Job> jobs; | 3551 core.List<Job> jobs; |
3343 /** Set if there may be more results than fit in this response. */ | 3552 /** Set if there may be more results than fit in this response. */ |
3344 core.String nextPageToken; | 3553 core.String nextPageToken; |
3345 | 3554 |
(...skipping 20 matching lines...) Expand all Loading... |
3366 _json["jobs"] = jobs.map((value) => (value).toJson()).toList(); | 3575 _json["jobs"] = jobs.map((value) => (value).toJson()).toList(); |
3367 } | 3576 } |
3368 if (nextPageToken != null) { | 3577 if (nextPageToken != null) { |
3369 _json["nextPageToken"] = nextPageToken; | 3578 _json["nextPageToken"] = nextPageToken; |
3370 } | 3579 } |
3371 return _json; | 3580 return _json; |
3372 } | 3581 } |
3373 } | 3582 } |
3374 | 3583 |
3375 /** | 3584 /** |
3376 * MapTask consists of an ordered set of instructions, each of which describes | 3585 * MapTask consists of an ordered set of instructions, each of which |
3377 * one particular low-level operation for the worker to perform in order to | 3586 * describes one particular low-level operation for the worker to |
3378 * accomplish the MapTask's WorkItem. Each instruction must appear in the list | 3587 * perform in order to accomplish the MapTask's WorkItem. |
3379 * before any instructions which depends on its output. | 3588 * |
| 3589 * Each instruction must appear in the list before any instructions which |
| 3590 * depends on its output. |
3380 */ | 3591 */ |
3381 class MapTask { | 3592 class MapTask { |
3382 /** The instructions in the MapTask. */ | 3593 /** The instructions in the MapTask. */ |
3383 core.List<ParallelInstruction> instructions; | 3594 core.List<ParallelInstruction> instructions; |
3384 /** | 3595 /** |
3385 * System-defined name of the stage containing this MapTask. Unique across the | 3596 * System-defined name of the stage containing this MapTask. |
3386 * workflow. | 3597 * Unique across the workflow. |
3387 */ | 3598 */ |
3388 core.String stageName; | 3599 core.String stageName; |
3389 /** System-defined name of this MapTask. Unique across the workflow. */ | 3600 /** |
| 3601 * System-defined name of this MapTask. |
| 3602 * Unique across the workflow. |
| 3603 */ |
3390 core.String systemName; | 3604 core.String systemName; |
3391 | 3605 |
3392 MapTask(); | 3606 MapTask(); |
3393 | 3607 |
3394 MapTask.fromJson(core.Map _json) { | 3608 MapTask.fromJson(core.Map _json) { |
3395 if (_json.containsKey("instructions")) { | 3609 if (_json.containsKey("instructions")) { |
3396 instructions = _json["instructions"].map((value) => new ParallelInstructio
n.fromJson(value)).toList(); | 3610 instructions = _json["instructions"].map((value) => new ParallelInstructio
n.fromJson(value)).toList(); |
3397 } | 3611 } |
3398 if (_json.containsKey("stageName")) { | 3612 if (_json.containsKey("stageName")) { |
3399 stageName = _json["stageName"]; | 3613 stageName = _json["stageName"]; |
(...skipping 17 matching lines...) Expand all Loading... |
3417 return _json; | 3631 return _json; |
3418 } | 3632 } |
3419 } | 3633 } |
3420 | 3634 |
3421 /** | 3635 /** |
3422 * The metric short id is returned to the user alongside an offset into | 3636 * The metric short id is returned to the user alongside an offset into |
3423 * ReportWorkItemStatusRequest | 3637 * ReportWorkItemStatusRequest |
3424 */ | 3638 */ |
3425 class MetricShortId { | 3639 class MetricShortId { |
3426 /** | 3640 /** |
3427 * The index of the corresponding metric in the ReportWorkItemStatusRequest. | 3641 * The index of the corresponding metric in |
3428 * Required. | 3642 * the ReportWorkItemStatusRequest. Required. |
3429 */ | 3643 */ |
3430 core.int metricIndex; | 3644 core.int metricIndex; |
3431 /** The service-generated short identifier for the metric. */ | 3645 /** The service-generated short identifier for the metric. */ |
3432 core.String shortId; | 3646 core.String shortId; |
3433 | 3647 |
3434 MetricShortId(); | 3648 MetricShortId(); |
3435 | 3649 |
3436 MetricShortId.fromJson(core.Map _json) { | 3650 MetricShortId.fromJson(core.Map _json) { |
3437 if (_json.containsKey("metricIndex")) { | 3651 if (_json.containsKey("metricIndex")) { |
3438 metricIndex = _json["metricIndex"]; | 3652 metricIndex = _json["metricIndex"]; |
3439 } | 3653 } |
3440 if (_json.containsKey("shortId")) { | 3654 if (_json.containsKey("shortId")) { |
3441 shortId = _json["shortId"]; | 3655 shortId = _json["shortId"]; |
3442 } | 3656 } |
3443 } | 3657 } |
3444 | 3658 |
3445 core.Map toJson() { | 3659 core.Map toJson() { |
3446 var _json = new core.Map(); | 3660 var _json = new core.Map(); |
3447 if (metricIndex != null) { | 3661 if (metricIndex != null) { |
3448 _json["metricIndex"] = metricIndex; | 3662 _json["metricIndex"] = metricIndex; |
3449 } | 3663 } |
3450 if (shortId != null) { | 3664 if (shortId != null) { |
3451 _json["shortId"] = shortId; | 3665 _json["shortId"] = shortId; |
3452 } | 3666 } |
3453 return _json; | 3667 return _json; |
3454 } | 3668 } |
3455 } | 3669 } |
3456 | 3670 |
3457 /** | 3671 /** |
3458 * Identifies a metric, by describing the source which generated the metric. | 3672 * Identifies a metric, by describing the source which generated the |
| 3673 * metric. |
3459 */ | 3674 */ |
3460 class MetricStructuredName { | 3675 class MetricStructuredName { |
3461 /** | 3676 /** |
3462 * Zero or more labeled fields which identify the part of the job this metric | 3677 * Zero or more labeled fields which identify the part of the job this |
3463 * is associated with, such as the name of a step or collection. For example, | 3678 * metric is associated with, such as the name of a step or collection. |
3464 * built-in counters associated with steps will have context['step'] = . | 3679 * |
3465 * Counters associated with PCollections in the SDK will have | 3680 * For example, built-in counters associated with steps will have |
3466 * context['pcollection'] = | 3681 * context['step'] = <step-name>. Counters associated with PCollections |
3467 * . | 3682 * in the SDK will have context['pcollection'] = <pcollection-name>. |
3468 */ | 3683 */ |
3469 core.Map<core.String, core.String> context; | 3684 core.Map<core.String, core.String> context; |
3470 /** Worker-defined metric name. */ | 3685 /** Worker-defined metric name. */ |
3471 core.String name; | 3686 core.String name; |
3472 /** | 3687 /** |
3473 * Origin (namespace) of metric name. May be blank for user-define metrics; | 3688 * Origin (namespace) of metric name. May be blank for user-define metrics; |
3474 * will be "dataflow" for metrics defined by the Dataflow service or SDK. | 3689 * will be "dataflow" for metrics defined by the Dataflow service or SDK. |
3475 */ | 3690 */ |
3476 core.String origin; | 3691 core.String origin; |
3477 | 3692 |
(...skipping 22 matching lines...) Expand all Loading... |
3500 if (origin != null) { | 3715 if (origin != null) { |
3501 _json["origin"] = origin; | 3716 _json["origin"] = origin; |
3502 } | 3717 } |
3503 return _json; | 3718 return _json; |
3504 } | 3719 } |
3505 } | 3720 } |
3506 | 3721 |
3507 /** Describes the state of a metric. */ | 3722 /** Describes the state of a metric. */ |
3508 class MetricUpdate { | 3723 class MetricUpdate { |
3509 /** | 3724 /** |
3510 * True if this metric is reported as the total cumulative aggregate value | 3725 * True if this metric is reported as the total cumulative aggregate |
3511 * accumulated since the worker started working on this WorkItem. By default | 3726 * value accumulated since the worker started working on this WorkItem. |
3512 * this is false, indicating that this metric is reported as a delta that is | 3727 * By default this is false, indicating that this metric is reported |
3513 * not associated with any WorkItem. | 3728 * as a delta that is not associated with any WorkItem. |
3514 */ | 3729 */ |
3515 core.bool cumulative; | 3730 core.bool cumulative; |
3516 /** | 3731 /** |
3517 * Worker-computed aggregate value for internal use by the Dataflow service. | 3732 * Worker-computed aggregate value for internal use by the Dataflow |
| 3733 * service. |
3518 * | 3734 * |
3519 * The values for Object must be JSON objects. It can consist of `num`, | 3735 * The values for Object must be JSON objects. It can consist of `num`, |
3520 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3736 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3521 */ | 3737 */ |
3522 core.Object internal; | 3738 core.Object internal; |
3523 /** | 3739 /** |
3524 * Metric aggregation kind. The possible metric aggregation kinds are "Sum", | 3740 * Metric aggregation kind. The possible metric aggregation kinds are |
3525 * "Max", "Min", "Mean", "Set", "And", and "Or". The specified aggregation | 3741 * "Sum", "Max", "Min", "Mean", "Set", "And", and "Or". |
3526 * kind is case-insensitive. If omitted, this is not an aggregated value but | 3742 * The specified aggregation kind is case-insensitive. |
3527 * instead a single metric sample value. | 3743 * |
| 3744 * If omitted, this is not an aggregated value but instead |
| 3745 * a single metric sample value. |
3528 */ | 3746 */ |
3529 core.String kind; | 3747 core.String kind; |
3530 /** | 3748 /** |
3531 * Worker-computed aggregate value for the "Mean" aggregation kind. This holds | 3749 * Worker-computed aggregate value for the "Mean" aggregation kind. |
3532 * the count of the aggregated values and is used in combination with mean_sum | 3750 * This holds the count of the aggregated values and is used in combination |
3533 * above to obtain the actual mean aggregate value. The only possible value | 3751 * with mean_sum above to obtain the actual mean aggregate value. |
3534 * type is Long. | 3752 * The only possible value type is Long. |
3535 * | 3753 * |
3536 * The values for Object must be JSON objects. It can consist of `num`, | 3754 * The values for Object must be JSON objects. It can consist of `num`, |
3537 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3755 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3538 */ | 3756 */ |
3539 core.Object meanCount; | 3757 core.Object meanCount; |
3540 /** | 3758 /** |
3541 * Worker-computed aggregate value for the "Mean" aggregation kind. This holds | 3759 * Worker-computed aggregate value for the "Mean" aggregation kind. |
3542 * the sum of the aggregated values and is used in combination with mean_count | 3760 * This holds the sum of the aggregated values and is used in combination |
3543 * below to obtain the actual mean aggregate value. The only possible value | 3761 * with mean_count below to obtain the actual mean aggregate value. |
3544 * types are Long and Double. | 3762 * The only possible value types are Long and Double. |
3545 * | 3763 * |
3546 * The values for Object must be JSON objects. It can consist of `num`, | 3764 * The values for Object must be JSON objects. It can consist of `num`, |
3547 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3765 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3548 */ | 3766 */ |
3549 core.Object meanSum; | 3767 core.Object meanSum; |
3550 /** Name of the metric. */ | 3768 /** Name of the metric. */ |
3551 MetricStructuredName name; | 3769 MetricStructuredName name; |
3552 /** | 3770 /** |
3553 * Worker-computed aggregate value for aggregation kinds "Sum", "Max", "Min", | 3771 * Worker-computed aggregate value for aggregation kinds "Sum", "Max", "Min", |
3554 * "And", and "Or". The possible value types are Long, Double, and Boolean. | 3772 * "And", and "Or". The possible value types are Long, Double, and Boolean. |
3555 * | 3773 * |
3556 * The values for Object must be JSON objects. It can consist of `num`, | 3774 * The values for Object must be JSON objects. It can consist of `num`, |
3557 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3775 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3558 */ | 3776 */ |
3559 core.Object scalar; | 3777 core.Object scalar; |
3560 /** | 3778 /** |
3561 * Worker-computed aggregate value for the "Set" aggregation kind. The only | 3779 * Worker-computed aggregate value for the "Set" aggregation kind. The only |
3562 * possible value type is a list of Values whose type can be Long, Double, or | 3780 * possible value type is a list of Values whose type can be Long, Double, |
3563 * String, according to the metric's type. All Values in the list must be of | 3781 * or String, according to the metric's type. All Values in the list must |
3564 * the same type. | 3782 * be of the same type. |
3565 * | 3783 * |
3566 * The values for Object must be JSON objects. It can consist of `num`, | 3784 * The values for Object must be JSON objects. It can consist of `num`, |
3567 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3785 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3568 */ | 3786 */ |
3569 core.Object set; | 3787 core.Object set; |
3570 /** | 3788 /** |
3571 * Timestamp associated with the metric value. Optional when workers are | 3789 * Timestamp associated with the metric value. Optional when workers are |
3572 * reporting work progress; it will be filled in responses from the metrics | 3790 * reporting work progress; it will be filled in responses from the |
3573 * API. | 3791 * metrics API. |
3574 */ | 3792 */ |
3575 core.String updateTime; | 3793 core.String updateTime; |
3576 | 3794 |
3577 MetricUpdate(); | 3795 MetricUpdate(); |
3578 | 3796 |
3579 MetricUpdate.fromJson(core.Map _json) { | 3797 MetricUpdate.fromJson(core.Map _json) { |
3580 if (_json.containsKey("cumulative")) { | 3798 if (_json.containsKey("cumulative")) { |
3581 cumulative = _json["cumulative"]; | 3799 cumulative = _json["cumulative"]; |
3582 } | 3800 } |
3583 if (_json.containsKey("internal")) { | 3801 if (_json.containsKey("internal")) { |
(...skipping 51 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3635 if (updateTime != null) { | 3853 if (updateTime != null) { |
3636 _json["updateTime"] = updateTime; | 3854 _json["updateTime"] = updateTime; |
3637 } | 3855 } |
3638 return _json; | 3856 return _json; |
3639 } | 3857 } |
3640 } | 3858 } |
3641 | 3859 |
3642 /** Describes mounted data disk. */ | 3860 /** Describes mounted data disk. */ |
3643 class MountedDataDisk { | 3861 class MountedDataDisk { |
3644 /** | 3862 /** |
3645 * The name of the data disk. This name is local to the Google Cloud Platform | 3863 * The name of the data disk. |
3646 * project and uniquely identifies the disk within that project, for example | 3864 * This name is local to the Google Cloud Platform project and uniquely |
| 3865 * identifies the disk within that project, for example |
3647 * "myproject-1014-104817-4c2-harness-0-disk-1". | 3866 * "myproject-1014-104817-4c2-harness-0-disk-1". |
3648 */ | 3867 */ |
3649 core.String dataDisk; | 3868 core.String dataDisk; |
3650 | 3869 |
3651 MountedDataDisk(); | 3870 MountedDataDisk(); |
3652 | 3871 |
3653 MountedDataDisk.fromJson(core.Map _json) { | 3872 MountedDataDisk.fromJson(core.Map _json) { |
3654 if (_json.containsKey("dataDisk")) { | 3873 if (_json.containsKey("dataDisk")) { |
3655 dataDisk = _json["dataDisk"]; | 3874 dataDisk = _json["dataDisk"]; |
3656 } | 3875 } |
3657 } | 3876 } |
3658 | 3877 |
3659 core.Map toJson() { | 3878 core.Map toJson() { |
3660 var _json = new core.Map(); | 3879 var _json = new core.Map(); |
3661 if (dataDisk != null) { | 3880 if (dataDisk != null) { |
3662 _json["dataDisk"] = dataDisk; | 3881 _json["dataDisk"] = dataDisk; |
3663 } | 3882 } |
3664 return _json; | 3883 return _json; |
3665 } | 3884 } |
3666 } | 3885 } |
3667 | 3886 |
3668 /** Information about an output of a multi-output DoFn. */ | 3887 /** Information about an output of a multi-output DoFn. */ |
3669 class MultiOutputInfo { | 3888 class MultiOutputInfo { |
3670 /** | 3889 /** |
3671 * The id of the tag the user code will emit to this output by; this should | 3890 * The id of the tag the user code will emit to this output by; this |
3672 * correspond to the tag of some SideInputInfo. | 3891 * should correspond to the tag of some SideInputInfo. |
3673 */ | 3892 */ |
3674 core.String tag; | 3893 core.String tag; |
3675 | 3894 |
3676 MultiOutputInfo(); | 3895 MultiOutputInfo(); |
3677 | 3896 |
3678 MultiOutputInfo.fromJson(core.Map _json) { | 3897 MultiOutputInfo.fromJson(core.Map _json) { |
3679 if (_json.containsKey("tag")) { | 3898 if (_json.containsKey("tag")) { |
3680 tag = _json["tag"]; | 3899 tag = _json["tag"]; |
3681 } | 3900 } |
3682 } | 3901 } |
3683 | 3902 |
3684 core.Map toJson() { | 3903 core.Map toJson() { |
3685 var _json = new core.Map(); | 3904 var _json = new core.Map(); |
3686 if (tag != null) { | 3905 if (tag != null) { |
3687 _json["tag"] = tag; | 3906 _json["tag"] = tag; |
3688 } | 3907 } |
3689 return _json; | 3908 return _json; |
3690 } | 3909 } |
3691 } | 3910 } |
3692 | 3911 |
3693 /** Basic metadata about a counter. */ | 3912 /** Basic metadata about a counter. */ |
3694 class NameAndKind { | 3913 class NameAndKind { |
3695 /** | 3914 /** |
3696 * Counter aggregation kind. | 3915 * Counter aggregation kind. |
3697 * Possible string values are: | 3916 * Possible string values are: |
3698 * - "INVALID" : A INVALID. | 3917 * - "INVALID" : Counter aggregation kind was not set. |
3699 * - "SUM" : A SUM. | 3918 * - "SUM" : Aggregated value is the sum of all contributed values. |
3700 * - "MAX" : A MAX. | 3919 * - "MAX" : Aggregated value is the max of all contributed values. |
3701 * - "MIN" : A MIN. | 3920 * - "MIN" : Aggregated value is the min of all contributed values. |
3702 * - "MEAN" : A MEAN. | 3921 * - "MEAN" : Aggregated value is the mean of all contributed values. |
3703 * - "OR" : A OR. | 3922 * - "OR" : Aggregated value represents the logical 'or' of all contributed |
3704 * - "AND" : A AND. | 3923 * values. |
3705 * - "SET" : A SET. | 3924 * - "AND" : Aggregated value represents the logical 'and' of all contributed |
3706 * - "DISTRIBUTION" : A DISTRIBUTION. | 3925 * values. |
| 3926 * - "SET" : Aggregated value is a set of unique contributed values. |
| 3927 * - "DISTRIBUTION" : Aggregated value captures statistics about a |
| 3928 * distribution. |
3707 */ | 3929 */ |
3708 core.String kind; | 3930 core.String kind; |
3709 /** Name of the counter. */ | 3931 /** Name of the counter. */ |
3710 core.String name; | 3932 core.String name; |
3711 | 3933 |
3712 NameAndKind(); | 3934 NameAndKind(); |
3713 | 3935 |
3714 NameAndKind.fromJson(core.Map _json) { | 3936 NameAndKind.fromJson(core.Map _json) { |
3715 if (_json.containsKey("kind")) { | 3937 if (_json.containsKey("kind")) { |
3716 kind = _json["kind"]; | 3938 kind = _json["kind"]; |
3717 } | 3939 } |
3718 if (_json.containsKey("name")) { | 3940 if (_json.containsKey("name")) { |
3719 name = _json["name"]; | 3941 name = _json["name"]; |
3720 } | 3942 } |
3721 } | 3943 } |
3722 | 3944 |
3723 core.Map toJson() { | 3945 core.Map toJson() { |
3724 var _json = new core.Map(); | 3946 var _json = new core.Map(); |
3725 if (kind != null) { | 3947 if (kind != null) { |
3726 _json["kind"] = kind; | 3948 _json["kind"] = kind; |
3727 } | 3949 } |
3728 if (name != null) { | 3950 if (name != null) { |
3729 _json["name"] = name; | 3951 _json["name"] = name; |
3730 } | 3952 } |
3731 return _json; | 3953 return _json; |
3732 } | 3954 } |
3733 } | 3955 } |
3734 | 3956 |
3735 /** | 3957 /** |
3736 * The packages that must be installed in order for a worker to run the steps of | 3958 * The packages that must be installed in order for a worker to run the |
3737 * the Cloud Dataflow job that will be assigned to its worker pool. This is the | 3959 * steps of the Cloud Dataflow job that will be assigned to its worker |
3738 * mechanism by which the Cloud Dataflow SDK causes code to be loaded onto the | 3960 * pool. |
3739 * workers. For example, the Cloud Dataflow Java SDK might use this to install | 3961 * |
3740 * jars containing the user's code and all of the various dependencies | 3962 * This is the mechanism by which the Cloud Dataflow SDK causes code to |
3741 * (libraries, data files, etc.) required in order for that code to run. | 3963 * be loaded onto the workers. For example, the Cloud Dataflow Java SDK |
| 3964 * might use this to install jars containing the user's code and all of the |
| 3965 * various dependencies (libraries, data files, etc.) required in order |
| 3966 * for that code to run. |
3742 */ | 3967 */ |
3743 class Package { | 3968 class Package { |
3744 /** | 3969 /** |
3745 * The resource to read the package from. The supported resource type is: | 3970 * The resource to read the package from. The supported resource type is: |
3746 * Google Cloud Storage: storage.googleapis.com/{bucket} | 3971 * |
3747 * bucket.storage.googleapis.com/ | 3972 * Google Cloud Storage: |
| 3973 * |
| 3974 * storage.googleapis.com/{bucket} |
| 3975 * bucket.storage.googleapis.com/ |
3748 */ | 3976 */ |
3749 core.String location; | 3977 core.String location; |
3750 /** The name of the package. */ | 3978 /** The name of the package. */ |
3751 core.String name; | 3979 core.String name; |
3752 | 3980 |
3753 Package(); | 3981 Package(); |
3754 | 3982 |
3755 Package.fromJson(core.Map _json) { | 3983 Package.fromJson(core.Map _json) { |
3756 if (_json.containsKey("location")) { | 3984 if (_json.containsKey("location")) { |
3757 location = _json["location"]; | 3985 location = _json["location"]; |
3758 } | 3986 } |
3759 if (_json.containsKey("name")) { | 3987 if (_json.containsKey("name")) { |
3760 name = _json["name"]; | 3988 name = _json["name"]; |
3761 } | 3989 } |
3762 } | 3990 } |
3763 | 3991 |
3764 core.Map toJson() { | 3992 core.Map toJson() { |
3765 var _json = new core.Map(); | 3993 var _json = new core.Map(); |
3766 if (location != null) { | 3994 if (location != null) { |
3767 _json["location"] = location; | 3995 _json["location"] = location; |
3768 } | 3996 } |
3769 if (name != null) { | 3997 if (name != null) { |
3770 _json["name"] = name; | 3998 _json["name"] = name; |
3771 } | 3999 } |
3772 return _json; | 4000 return _json; |
3773 } | 4001 } |
3774 } | 4002 } |
3775 | 4003 |
3776 /** | 4004 /** |
3777 * An instruction that does a ParDo operation. Takes one main input and zero or | 4005 * An instruction that does a ParDo operation. |
3778 * more side inputs, and produces zero or more outputs. Runs user code. | 4006 * Takes one main input and zero or more side inputs, and produces |
| 4007 * zero or more outputs. |
| 4008 * Runs user code. |
3779 */ | 4009 */ |
3780 class ParDoInstruction { | 4010 class ParDoInstruction { |
3781 /** The input. */ | 4011 /** The input. */ |
3782 InstructionInput input; | 4012 InstructionInput input; |
3783 /** Information about each of the outputs, if user_fn is a MultiDoFn. */ | 4013 /** Information about each of the outputs, if user_fn is a MultiDoFn. */ |
3784 core.List<MultiOutputInfo> multiOutputInfos; | 4014 core.List<MultiOutputInfo> multiOutputInfos; |
3785 /** The number of outputs. */ | 4015 /** The number of outputs. */ |
3786 core.int numOutputs; | 4016 core.int numOutputs; |
3787 /** Zero or more side inputs. */ | 4017 /** Zero or more side inputs. */ |
3788 core.List<SideInputInfo> sideInputs; | 4018 core.List<SideInputInfo> sideInputs; |
3789 /** | 4019 /** |
3790 * The user function to invoke. | 4020 * The user function to invoke. |
3791 * | 4021 * |
3792 * The values for Object must be JSON objects. It can consist of `num`, | 4022 * The values for Object must be JSON objects. It can consist of `num`, |
3793 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4023 * `String`, `bool` and `null` as well as `Map` and `List` values. |
(...skipping 50 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3844 /** System-defined name for the operation in the original workflow graph. */ | 4074 /** System-defined name for the operation in the original workflow graph. */ |
3845 core.String originalName; | 4075 core.String originalName; |
3846 /** Describes the outputs of the instruction. */ | 4076 /** Describes the outputs of the instruction. */ |
3847 core.List<InstructionOutput> outputs; | 4077 core.List<InstructionOutput> outputs; |
3848 /** Additional information for ParDo instructions. */ | 4078 /** Additional information for ParDo instructions. */ |
3849 ParDoInstruction parDo; | 4079 ParDoInstruction parDo; |
3850 /** Additional information for PartialGroupByKey instructions. */ | 4080 /** Additional information for PartialGroupByKey instructions. */ |
3851 PartialGroupByKeyInstruction partialGroupByKey; | 4081 PartialGroupByKeyInstruction partialGroupByKey; |
3852 /** Additional information for Read instructions. */ | 4082 /** Additional information for Read instructions. */ |
3853 ReadInstruction read; | 4083 ReadInstruction read; |
3854 /** System-defined name of this operation. Unique across the workflow. */ | 4084 /** |
| 4085 * System-defined name of this operation. |
| 4086 * Unique across the workflow. |
| 4087 */ |
3855 core.String systemName; | 4088 core.String systemName; |
3856 /** Additional information for Write instructions. */ | 4089 /** Additional information for Write instructions. */ |
3857 WriteInstruction write; | 4090 WriteInstruction write; |
3858 | 4091 |
3859 ParallelInstruction(); | 4092 ParallelInstruction(); |
3860 | 4093 |
3861 ParallelInstruction.fromJson(core.Map _json) { | 4094 ParallelInstruction.fromJson(core.Map _json) { |
3862 if (_json.containsKey("flatten")) { | 4095 if (_json.containsKey("flatten")) { |
3863 flatten = new FlattenInstruction.fromJson(_json["flatten"]); | 4096 flatten = new FlattenInstruction.fromJson(_json["flatten"]); |
3864 } | 4097 } |
(...skipping 50 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3915 _json["systemName"] = systemName; | 4148 _json["systemName"] = systemName; |
3916 } | 4149 } |
3917 if (write != null) { | 4150 if (write != null) { |
3918 _json["write"] = (write).toJson(); | 4151 _json["write"] = (write).toJson(); |
3919 } | 4152 } |
3920 return _json; | 4153 return _json; |
3921 } | 4154 } |
3922 } | 4155 } |
3923 | 4156 |
3924 /** | 4157 /** |
3925 * An instruction that does a partial group-by-key. One input and one output. | 4158 * An instruction that does a partial group-by-key. |
| 4159 * One input and one output. |
3926 */ | 4160 */ |
3927 class PartialGroupByKeyInstruction { | 4161 class PartialGroupByKeyInstruction { |
3928 /** Describes the input to the partial group-by-key instruction. */ | 4162 /** Describes the input to the partial group-by-key instruction. */ |
3929 InstructionInput input; | 4163 InstructionInput input; |
3930 /** | 4164 /** |
3931 * The codec to use for interpreting an element in the input PTable. | 4165 * The codec to use for interpreting an element in the input PTable. |
3932 * | 4166 * |
3933 * The values for Object must be JSON objects. It can consist of `num`, | 4167 * The values for Object must be JSON objects. It can consist of `num`, |
3934 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4168 * `String`, `bool` and `null` as well as `Map` and `List` values. |
3935 */ | 4169 */ |
(...skipping 59 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
3995 _json["sideInputs"] = sideInputs.map((value) => (value).toJson()).toList()
; | 4229 _json["sideInputs"] = sideInputs.map((value) => (value).toJson()).toList()
; |
3996 } | 4230 } |
3997 if (valueCombiningFn != null) { | 4231 if (valueCombiningFn != null) { |
3998 _json["valueCombiningFn"] = valueCombiningFn; | 4232 _json["valueCombiningFn"] = valueCombiningFn; |
3999 } | 4233 } |
4000 return _json; | 4234 return _json; |
4001 } | 4235 } |
4002 } | 4236 } |
4003 | 4237 |
4004 /** | 4238 /** |
4005 * Position defines a position within a collection of data. The value can be | 4239 * Position defines a position within a collection of data. The value |
4006 * either the end position, a key (used with ordered collections), a byte | 4240 * can be either the end position, a key (used with ordered |
4007 * offset, or a record index. | 4241 * collections), a byte offset, or a record index. |
4008 */ | 4242 */ |
4009 class Position { | 4243 class Position { |
4010 /** Position is a byte offset. */ | 4244 /** Position is a byte offset. */ |
4011 core.String byteOffset; | 4245 core.String byteOffset; |
4012 /** CloudPosition is a concat position. */ | 4246 /** CloudPosition is a concat position. */ |
4013 ConcatPosition concatPosition; | 4247 ConcatPosition concatPosition; |
4014 /** | 4248 /** |
4015 * Position is past all other positions. Also useful for the end position of | 4249 * Position is past all other positions. Also useful for the end |
4016 * an unbounded range. | 4250 * position of an unbounded range. |
4017 */ | 4251 */ |
4018 core.bool end; | 4252 core.bool end; |
4019 /** Position is a string key, ordered lexicographically. */ | 4253 /** Position is a string key, ordered lexicographically. */ |
4020 core.String key; | 4254 core.String key; |
4021 /** Position is a record index. */ | 4255 /** Position is a record index. */ |
4022 core.String recordIndex; | 4256 core.String recordIndex; |
4023 /** | 4257 /** |
4024 * CloudPosition is a base64 encoded BatchShufflePosition (with FIXED | 4258 * CloudPosition is a base64 encoded BatchShufflePosition (with FIXED |
4025 * sharding). | 4259 * sharding). |
4026 */ | 4260 */ |
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4067 _json["recordIndex"] = recordIndex; | 4301 _json["recordIndex"] = recordIndex; |
4068 } | 4302 } |
4069 if (shufflePosition != null) { | 4303 if (shufflePosition != null) { |
4070 _json["shufflePosition"] = shufflePosition; | 4304 _json["shufflePosition"] = shufflePosition; |
4071 } | 4305 } |
4072 return _json; | 4306 return _json; |
4073 } | 4307 } |
4074 } | 4308 } |
4075 | 4309 |
4076 /** | 4310 /** |
4077 * Identifies a pubsub location to use for transferring data into or out of a | 4311 * Identifies a pubsub location to use for transferring data into or |
4078 * streaming Dataflow job. | 4312 * out of a streaming Dataflow job. |
4079 */ | 4313 */ |
4080 class PubsubLocation { | 4314 class PubsubLocation { |
4081 /** Indicates whether the pipeline allows late-arriving data. */ | 4315 /** Indicates whether the pipeline allows late-arriving data. */ |
4082 core.bool dropLateData; | 4316 core.bool dropLateData; |
4083 /** | 4317 /** |
4084 * If set, contains a pubsub label from which to extract record ids. If left | 4318 * If set, contains a pubsub label from which to extract record ids. |
4085 * empty, record deduplication will be strictly best effort. | 4319 * If left empty, record deduplication will be strictly best effort. |
4086 */ | 4320 */ |
4087 core.String idLabel; | 4321 core.String idLabel; |
4088 /** | 4322 /** |
4089 * A pubsub subscription, in the form of "pubsub.googleapis.com/subscriptions/ | 4323 * A pubsub subscription, in the form of |
4090 * /" | 4324 * "pubsub.googleapis.com/subscriptions/<project-id>/<subscription-name>" |
4091 */ | 4325 */ |
4092 core.String subscription; | 4326 core.String subscription; |
4093 /** | 4327 /** |
4094 * If set, contains a pubsub label from which to extract record timestamps. If | 4328 * If set, contains a pubsub label from which to extract record timestamps. |
4095 * left empty, record timestamps will be generated upon arrival. | 4329 * If left empty, record timestamps will be generated upon arrival. |
4096 */ | 4330 */ |
4097 core.String timestampLabel; | 4331 core.String timestampLabel; |
4098 /** | 4332 /** |
4099 * A pubsub topic, in the form of "pubsub.googleapis.com/topics/ | 4333 * A pubsub topic, in the form of |
4100 * /" | 4334 * "pubsub.googleapis.com/topics/<project-id>/<topic-name>" |
4101 */ | 4335 */ |
4102 core.String topic; | 4336 core.String topic; |
4103 /** | 4337 /** |
4104 * If set, specifies the pubsub subscription that will be used for tracking | 4338 * If set, specifies the pubsub subscription that will be used for tracking |
4105 * custom time timestamps for watermark estimation. | 4339 * custom time timestamps for watermark estimation. |
4106 */ | 4340 */ |
4107 core.String trackingSubscription; | 4341 core.String trackingSubscription; |
4108 /** If true, then the client has requested to get pubsub attributes. */ | 4342 /** If true, then the client has requested to get pubsub attributes. */ |
4109 core.bool withAttributes; | 4343 core.bool withAttributes; |
4110 | 4344 |
(...skipping 43 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4154 if (trackingSubscription != null) { | 4388 if (trackingSubscription != null) { |
4155 _json["trackingSubscription"] = trackingSubscription; | 4389 _json["trackingSubscription"] = trackingSubscription; |
4156 } | 4390 } |
4157 if (withAttributes != null) { | 4391 if (withAttributes != null) { |
4158 _json["withAttributes"] = withAttributes; | 4392 _json["withAttributes"] = withAttributes; |
4159 } | 4393 } |
4160 return _json; | 4394 return _json; |
4161 } | 4395 } |
4162 } | 4396 } |
4163 | 4397 |
4164 /** An instruction that reads records. Takes no inputs, produces one output. */ | 4398 /** |
| 4399 * An instruction that reads records. |
| 4400 * Takes no inputs, produces one output. |
| 4401 */ |
4165 class ReadInstruction { | 4402 class ReadInstruction { |
4166 /** The source to read from. */ | 4403 /** The source to read from. */ |
4167 Source source; | 4404 Source source; |
4168 | 4405 |
4169 ReadInstruction(); | 4406 ReadInstruction(); |
4170 | 4407 |
4171 ReadInstruction.fromJson(core.Map _json) { | 4408 ReadInstruction.fromJson(core.Map _json) { |
4172 if (_json.containsKey("source")) { | 4409 if (_json.containsKey("source")) { |
4173 source = new Source.fromJson(_json["source"]); | 4410 source = new Source.fromJson(_json["source"]); |
4174 } | 4411 } |
4175 } | 4412 } |
4176 | 4413 |
4177 core.Map toJson() { | 4414 core.Map toJson() { |
4178 var _json = new core.Map(); | 4415 var _json = new core.Map(); |
4179 if (source != null) { | 4416 if (source != null) { |
4180 _json["source"] = (source).toJson(); | 4417 _json["source"] = (source).toJson(); |
4181 } | 4418 } |
4182 return _json; | 4419 return _json; |
4183 } | 4420 } |
4184 } | 4421 } |
4185 | 4422 |
4186 /** Request to report the status of WorkItems. */ | 4423 /** Request to report the status of WorkItems. */ |
4187 class ReportWorkItemStatusRequest { | 4424 class ReportWorkItemStatusRequest { |
4188 /** The current timestamp at the worker. */ | 4425 /** The current timestamp at the worker. */ |
4189 core.String currentWorkerTime; | 4426 core.String currentWorkerTime; |
4190 /** The location which contains the WorkItem's job. */ | 4427 /** The location which contains the WorkItem's job. */ |
4191 core.String location; | 4428 core.String location; |
4192 /** | 4429 /** |
4193 * The order is unimportant, except that the order of the WorkItemServiceState | 4430 * The order is unimportant, except that the order of the |
4194 * messages in the ReportWorkItemStatusResponse corresponds to the order of | 4431 * WorkItemServiceState messages in the ReportWorkItemStatusResponse |
4195 * WorkItemStatus messages here. | 4432 * corresponds to the order of WorkItemStatus messages here. |
4196 */ | 4433 */ |
4197 core.List<WorkItemStatus> workItemStatuses; | 4434 core.List<WorkItemStatus> workItemStatuses; |
4198 /** | 4435 /** |
4199 * The ID of the worker reporting the WorkItem status. If this does not match | 4436 * The ID of the worker reporting the WorkItem status. If this |
4200 * the ID of the worker which the Dataflow service believes currently has the | 4437 * does not match the ID of the worker which the Dataflow service |
4201 * lease on the WorkItem, the report will be dropped (with an error response). | 4438 * believes currently has the lease on the WorkItem, the report |
| 4439 * will be dropped (with an error response). |
4202 */ | 4440 */ |
4203 core.String workerId; | 4441 core.String workerId; |
4204 | 4442 |
4205 ReportWorkItemStatusRequest(); | 4443 ReportWorkItemStatusRequest(); |
4206 | 4444 |
4207 ReportWorkItemStatusRequest.fromJson(core.Map _json) { | 4445 ReportWorkItemStatusRequest.fromJson(core.Map _json) { |
4208 if (_json.containsKey("currentWorkerTime")) { | 4446 if (_json.containsKey("currentWorkerTime")) { |
4209 currentWorkerTime = _json["currentWorkerTime"]; | 4447 currentWorkerTime = _json["currentWorkerTime"]; |
4210 } | 4448 } |
4211 if (_json.containsKey("location")) { | 4449 if (_json.containsKey("location")) { |
(...skipping 21 matching lines...) Expand all Loading... |
4233 if (workerId != null) { | 4471 if (workerId != null) { |
4234 _json["workerId"] = workerId; | 4472 _json["workerId"] = workerId; |
4235 } | 4473 } |
4236 return _json; | 4474 return _json; |
4237 } | 4475 } |
4238 } | 4476 } |
4239 | 4477 |
4240 /** Response from a request to report the status of WorkItems. */ | 4478 /** Response from a request to report the status of WorkItems. */ |
4241 class ReportWorkItemStatusResponse { | 4479 class ReportWorkItemStatusResponse { |
4242 /** | 4480 /** |
4243 * A set of messages indicating the service-side state for each WorkItem whose | 4481 * A set of messages indicating the service-side state for each |
4244 * status was reported, in the same order as the WorkItemStatus messages in | 4482 * WorkItem whose status was reported, in the same order as the |
4245 * the ReportWorkItemStatusRequest which resulting in this response. | 4483 * WorkItemStatus messages in the ReportWorkItemStatusRequest which |
| 4484 * resulting in this response. |
4246 */ | 4485 */ |
4247 core.List<WorkItemServiceState> workItemServiceStates; | 4486 core.List<WorkItemServiceState> workItemServiceStates; |
4248 | 4487 |
4249 ReportWorkItemStatusResponse(); | 4488 ReportWorkItemStatusResponse(); |
4250 | 4489 |
4251 ReportWorkItemStatusResponse.fromJson(core.Map _json) { | 4490 ReportWorkItemStatusResponse.fromJson(core.Map _json) { |
4252 if (_json.containsKey("workItemServiceStates")) { | 4491 if (_json.containsKey("workItemServiceStates")) { |
4253 workItemServiceStates = _json["workItemServiceStates"].map((value) => new
WorkItemServiceState.fromJson(value)).toList(); | 4492 workItemServiceStates = _json["workItemServiceStates"].map((value) => new
WorkItemServiceState.fromJson(value)).toList(); |
4254 } | 4493 } |
4255 } | 4494 } |
4256 | 4495 |
4257 core.Map toJson() { | 4496 core.Map toJson() { |
4258 var _json = new core.Map(); | 4497 var _json = new core.Map(); |
4259 if (workItemServiceStates != null) { | 4498 if (workItemServiceStates != null) { |
4260 _json["workItemServiceStates"] = workItemServiceStates.map((value) => (val
ue).toJson()).toList(); | 4499 _json["workItemServiceStates"] = workItemServiceStates.map((value) => (val
ue).toJson()).toList(); |
4261 } | 4500 } |
4262 return _json; | 4501 return _json; |
4263 } | 4502 } |
4264 } | 4503 } |
4265 | 4504 |
4266 /** | 4505 /** |
4267 * Represents the level of parallelism in a WorkItem's input, reported by the | 4506 * Represents the level of parallelism in a WorkItem's input, |
4268 * worker. | 4507 * reported by the worker. |
4269 */ | 4508 */ |
4270 class ReportedParallelism { | 4509 class ReportedParallelism { |
4271 /** | 4510 /** |
4272 * Specifies whether the parallelism is infinite. If true, "value" is ignored. | 4511 * Specifies whether the parallelism is infinite. If true, "value" is |
4273 * Infinite parallelism means the service will assume that the work item can | 4512 * ignored. |
4274 * always be split into more non-empty work items by dynamic splitting. This | 4513 * Infinite parallelism means the service will assume that the work item |
4275 * is a work-around for lack of support for infinity by the current JSON-based | 4514 * can always be split into more non-empty work items by dynamic splitting. |
4276 * Java RPC stack. | 4515 * This is a work-around for lack of support for infinity by the current |
| 4516 * JSON-based Java RPC stack. |
4277 */ | 4517 */ |
4278 core.bool isInfinite; | 4518 core.bool isInfinite; |
4279 /** Specifies the level of parallelism in case it is finite. */ | 4519 /** Specifies the level of parallelism in case it is finite. */ |
4280 core.double value; | 4520 core.double value; |
4281 | 4521 |
4282 ReportedParallelism(); | 4522 ReportedParallelism(); |
4283 | 4523 |
4284 ReportedParallelism.fromJson(core.Map _json) { | 4524 ReportedParallelism.fromJson(core.Map _json) { |
4285 if (_json.containsKey("isInfinite")) { | 4525 if (_json.containsKey("isInfinite")) { |
4286 isInfinite = _json["isInfinite"]; | 4526 isInfinite = _json["isInfinite"]; |
(...skipping 11 matching lines...) Expand all Loading... |
4298 if (value != null) { | 4538 if (value != null) { |
4299 _json["value"] = value; | 4539 _json["value"] = value; |
4300 } | 4540 } |
4301 return _json; | 4541 return _json; |
4302 } | 4542 } |
4303 } | 4543 } |
4304 | 4544 |
4305 /** | 4545 /** |
4306 * Worker metrics exported from workers. This contains resource utilization | 4546 * Worker metrics exported from workers. This contains resource utilization |
4307 * metrics accumulated from a variety of sources. For more information, see | 4547 * metrics accumulated from a variety of sources. For more information, see |
4308 * go/df-resource-signals. Note that this proto closely follows the structure of | 4548 * go/df-resource-signals. |
4309 * its DFE siblings in its contents. | |
4310 */ | 4549 */ |
4311 class ResourceUtilizationReport { | 4550 class ResourceUtilizationReport { |
4312 /** | 4551 /** CPU utilization samples. */ |
4313 * Each Struct must parallel DFE worker metrics protos (eg., cpu_time metric | 4552 core.List<CPUTime> cpuTime; |
4314 * will have nested values “timestamp_ms, total_ms, rate”). | |
4315 * | |
4316 * The values for Object must be JSON objects. It can consist of `num`, | |
4317 * `String`, `bool` and `null` as well as `Map` and `List` values. | |
4318 */ | |
4319 core.List<core.Map<core.String, core.Object>> metrics; | |
4320 | 4553 |
4321 ResourceUtilizationReport(); | 4554 ResourceUtilizationReport(); |
4322 | 4555 |
4323 ResourceUtilizationReport.fromJson(core.Map _json) { | 4556 ResourceUtilizationReport.fromJson(core.Map _json) { |
4324 if (_json.containsKey("metrics")) { | 4557 if (_json.containsKey("cpuTime")) { |
4325 metrics = _json["metrics"]; | 4558 cpuTime = _json["cpuTime"].map((value) => new CPUTime.fromJson(value)).toL
ist(); |
4326 } | 4559 } |
4327 } | 4560 } |
4328 | 4561 |
4329 core.Map toJson() { | 4562 core.Map toJson() { |
4330 var _json = new core.Map(); | 4563 var _json = new core.Map(); |
4331 if (metrics != null) { | 4564 if (cpuTime != null) { |
4332 _json["metrics"] = metrics; | 4565 _json["cpuTime"] = cpuTime.map((value) => (value).toJson()).toList(); |
4333 } | 4566 } |
4334 return _json; | 4567 return _json; |
4335 } | 4568 } |
4336 } | 4569 } |
4337 | 4570 |
4338 /** Service-side response to WorkerMessage reporting resource utilization. */ | 4571 /** Service-side response to WorkerMessage reporting resource utilization. */ |
4339 class ResourceUtilizationReportResponse { | 4572 class ResourceUtilizationReportResponse { |
4340 | 4573 |
4341 ResourceUtilizationReportResponse(); | 4574 ResourceUtilizationReportResponse(); |
4342 | 4575 |
4343 ResourceUtilizationReportResponse.fromJson(core.Map _json) { | 4576 ResourceUtilizationReportResponse.fromJson(core.Map _json) { |
4344 } | 4577 } |
4345 | 4578 |
4346 core.Map toJson() { | 4579 core.Map toJson() { |
4347 var _json = new core.Map(); | 4580 var _json = new core.Map(); |
4348 return _json; | 4581 return _json; |
4349 } | 4582 } |
4350 } | 4583 } |
4351 | 4584 |
4352 /** The environment values to set at runtime. */ | 4585 /** The environment values to set at runtime. */ |
4353 class RuntimeEnvironment { | 4586 class RuntimeEnvironment { |
4354 /** | 4587 /** |
4355 * Whether to bypass the safety checks for the job's temporary directory. Use | 4588 * Whether to bypass the safety checks for the job's temporary directory. |
4356 * with caution. | 4589 * Use with caution. |
4357 */ | 4590 */ |
4358 core.bool bypassTempDirValidation; | 4591 core.bool bypassTempDirValidation; |
4359 /** | 4592 /** |
4360 * The maximum number of Google Compute Engine instances to be made available | 4593 * The maximum number of Google Compute Engine instances to be made |
4361 * to your pipeline during execution, from 1 to 1000. | 4594 * available to your pipeline during execution, from 1 to 1000. |
4362 */ | 4595 */ |
4363 core.int maxWorkers; | 4596 core.int maxWorkers; |
4364 /** The email address of the service account to run the job as. */ | 4597 /** The email address of the service account to run the job as. */ |
4365 core.String serviceAccountEmail; | 4598 core.String serviceAccountEmail; |
4366 /** | 4599 /** |
4367 * The Cloud Storage path to use for temporary files. Must be a valid Cloud | 4600 * The Cloud Storage path to use for temporary files. |
4368 * Storage URL, beginning with `gs://`. | 4601 * Must be a valid Cloud Storage URL, beginning with `gs://`. |
4369 */ | 4602 */ |
4370 core.String tempLocation; | 4603 core.String tempLocation; |
4371 /** | 4604 /** |
4372 * The Compute Engine [availability | 4605 * The Compute Engine [availability |
4373 * zone](https://cloud.google.com/compute/docs/regions-zones/regions-zones) | 4606 * zone](https://cloud.google.com/compute/docs/regions-zones/regions-zones) |
4374 * for launching worker instances to run your pipeline. | 4607 * for launching worker instances to run your pipeline. |
4375 */ | 4608 */ |
4376 core.String zone; | 4609 core.String zone; |
4377 | 4610 |
4378 RuntimeEnvironment(); | 4611 RuntimeEnvironment(); |
(...skipping 68 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4447 if (data != null) { | 4680 if (data != null) { |
4448 _json["data"] = data; | 4681 _json["data"] = data; |
4449 } | 4682 } |
4450 if (workerId != null) { | 4683 if (workerId != null) { |
4451 _json["workerId"] = workerId; | 4684 _json["workerId"] = workerId; |
4452 } | 4685 } |
4453 return _json; | 4686 return _json; |
4454 } | 4687 } |
4455 } | 4688 } |
4456 | 4689 |
4457 /** Response to a send capture request. nothing */ | 4690 /** |
| 4691 * Response to a send capture request. |
| 4692 * nothing |
| 4693 */ |
4458 class SendDebugCaptureResponse { | 4694 class SendDebugCaptureResponse { |
4459 | 4695 |
4460 SendDebugCaptureResponse(); | 4696 SendDebugCaptureResponse(); |
4461 | 4697 |
4462 SendDebugCaptureResponse.fromJson(core.Map _json) { | 4698 SendDebugCaptureResponse.fromJson(core.Map _json) { |
4463 } | 4699 } |
4464 | 4700 |
4465 core.Map toJson() { | 4701 core.Map toJson() { |
4466 var _json = new core.Map(); | 4702 var _json = new core.Map(); |
4467 return _json; | 4703 return _json; |
(...skipping 46 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4514 | 4750 |
4515 /** Describes a particular function to invoke. */ | 4751 /** Describes a particular function to invoke. */ |
4516 class SeqMapTask { | 4752 class SeqMapTask { |
4517 /** Information about each of the inputs. */ | 4753 /** Information about each of the inputs. */ |
4518 core.List<SideInputInfo> inputs; | 4754 core.List<SideInputInfo> inputs; |
4519 /** The user-provided name of the SeqDo operation. */ | 4755 /** The user-provided name of the SeqDo operation. */ |
4520 core.String name; | 4756 core.String name; |
4521 /** Information about each of the outputs. */ | 4757 /** Information about each of the outputs. */ |
4522 core.List<SeqMapTaskOutputInfo> outputInfos; | 4758 core.List<SeqMapTaskOutputInfo> outputInfos; |
4523 /** | 4759 /** |
4524 * System-defined name of the stage containing the SeqDo operation. Unique | 4760 * System-defined name of the stage containing the SeqDo operation. |
4525 * across the workflow. | 4761 * Unique across the workflow. |
4526 */ | 4762 */ |
4527 core.String stageName; | 4763 core.String stageName; |
4528 /** | 4764 /** |
4529 * System-defined name of the SeqDo operation. Unique across the workflow. | 4765 * System-defined name of the SeqDo operation. |
| 4766 * Unique across the workflow. |
4530 */ | 4767 */ |
4531 core.String systemName; | 4768 core.String systemName; |
4532 /** | 4769 /** |
4533 * The user function to invoke. | 4770 * The user function to invoke. |
4534 * | 4771 * |
4535 * The values for Object must be JSON objects. It can consist of `num`, | 4772 * The values for Object must be JSON objects. It can consist of `num`, |
4536 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4773 * `String`, `bool` and `null` as well as `Map` and `List` values. |
4537 */ | 4774 */ |
4538 core.Map<core.String, core.Object> userFn; | 4775 core.Map<core.String, core.Object> userFn; |
4539 | 4776 |
(...skipping 108 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4648 class SideInputInfo { | 4885 class SideInputInfo { |
4649 /** | 4886 /** |
4650 * How to interpret the source element(s) as a side input value. | 4887 * How to interpret the source element(s) as a side input value. |
4651 * | 4888 * |
4652 * The values for Object must be JSON objects. It can consist of `num`, | 4889 * The values for Object must be JSON objects. It can consist of `num`, |
4653 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4890 * `String`, `bool` and `null` as well as `Map` and `List` values. |
4654 */ | 4891 */ |
4655 core.Map<core.String, core.Object> kind; | 4892 core.Map<core.String, core.Object> kind; |
4656 /** | 4893 /** |
4657 * The source(s) to read element(s) from to get the value of this side input. | 4894 * The source(s) to read element(s) from to get the value of this side input. |
4658 * If more than one source, then the elements are taken from the sources, in | 4895 * If more than one source, then the elements are taken from the |
4659 * the specified order if order matters. At least one source is required. | 4896 * sources, in the specified order if order matters. |
| 4897 * At least one source is required. |
4660 */ | 4898 */ |
4661 core.List<Source> sources; | 4899 core.List<Source> sources; |
4662 /** | 4900 /** |
4663 * The id of the tag the user code will access this side input by; this should | 4901 * The id of the tag the user code will access this side input by; |
4664 * correspond to the tag of some MultiOutputInfo. | 4902 * this should correspond to the tag of some MultiOutputInfo. |
4665 */ | 4903 */ |
4666 core.String tag; | 4904 core.String tag; |
4667 | 4905 |
4668 SideInputInfo(); | 4906 SideInputInfo(); |
4669 | 4907 |
4670 SideInputInfo.fromJson(core.Map _json) { | 4908 SideInputInfo.fromJson(core.Map _json) { |
4671 if (_json.containsKey("kind")) { | 4909 if (_json.containsKey("kind")) { |
4672 kind = _json["kind"]; | 4910 kind = _json["kind"]; |
4673 } | 4911 } |
4674 if (_json.containsKey("sources")) { | 4912 if (_json.containsKey("sources")) { |
(...skipping 55 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4730 if (spec != null) { | 4968 if (spec != null) { |
4731 _json["spec"] = spec; | 4969 _json["spec"] = spec; |
4732 } | 4970 } |
4733 return _json; | 4971 return _json; |
4734 } | 4972 } |
4735 } | 4973 } |
4736 | 4974 |
4737 /** A source that records can be read and decoded from. */ | 4975 /** A source that records can be read and decoded from. */ |
4738 class Source { | 4976 class Source { |
4739 /** | 4977 /** |
4740 * While splitting, sources may specify the produced bundles as differences | 4978 * While splitting, sources may specify the produced bundles |
4741 * against another source, in order to save backend-side memory and allow | 4979 * as differences against another source, in order to save backend-side |
4742 * bigger jobs. For details, see SourceSplitRequest. To support this use case, | 4980 * memory and allow bigger jobs. For details, see SourceSplitRequest. |
4743 * the full set of parameters of the source is logically obtained by taking | 4981 * To support this use case, the full set of parameters of the source |
4744 * the latest explicitly specified value of each parameter in the order: | 4982 * is logically obtained by taking the latest explicitly specified value |
| 4983 * of each parameter in the order: |
4745 * base_specs (later items win), spec (overrides anything in base_specs). | 4984 * base_specs (later items win), spec (overrides anything in base_specs). |
4746 * | 4985 * |
4747 * The values for Object must be JSON objects. It can consist of `num`, | 4986 * The values for Object must be JSON objects. It can consist of `num`, |
4748 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4987 * `String`, `bool` and `null` as well as `Map` and `List` values. |
4749 */ | 4988 */ |
4750 core.List<core.Map<core.String, core.Object>> baseSpecs; | 4989 core.List<core.Map<core.String, core.Object>> baseSpecs; |
4751 /** | 4990 /** |
4752 * The codec to use to decode data read from the source. | 4991 * The codec to use to decode data read from the source. |
4753 * | 4992 * |
4754 * The values for Object must be JSON objects. It can consist of `num`, | 4993 * The values for Object must be JSON objects. It can consist of `num`, |
4755 * `String`, `bool` and `null` as well as `Map` and `List` values. | 4994 * `String`, `bool` and `null` as well as `Map` and `List` values. |
4756 */ | 4995 */ |
4757 core.Map<core.String, core.Object> codec; | 4996 core.Map<core.String, core.Object> codec; |
4758 /** | 4997 /** |
4759 * Setting this value to true hints to the framework that the source doesn't | 4998 * Setting this value to true hints to the framework that the source |
4760 * need splitting, and using SourceSplitRequest on it would yield | 4999 * doesn't need splitting, and using SourceSplitRequest on it would |
4761 * SOURCE_SPLIT_OUTCOME_USE_CURRENT. E.g. a file splitter may set this to true | 5000 * yield SOURCE_SPLIT_OUTCOME_USE_CURRENT. |
4762 * when splitting a single file into a set of byte ranges of appropriate size, | 5001 * |
4763 * and set this to false when splitting a filepattern into individual files. | 5002 * E.g. a file splitter may set this to true when splitting a single file |
4764 * However, for efficiency, a file splitter may decide to produce file | 5003 * into a set of byte ranges of appropriate size, and set this |
4765 * subranges directly from the filepattern to avoid a splitting round-trip. | 5004 * to false when splitting a filepattern into individual files. |
4766 * See SourceSplitRequest for an overview of the splitting process. This field | 5005 * However, for efficiency, a file splitter may decide to produce |
4767 * is meaningful only in the Source objects populated by the user (e.g. when | 5006 * file subranges directly from the filepattern to avoid a splitting |
4768 * filling in a DerivedSource). Source objects supplied by the framework to | 5007 * round-trip. |
4769 * the user don't have this field populated. | 5008 * |
| 5009 * See SourceSplitRequest for an overview of the splitting process. |
| 5010 * |
| 5011 * This field is meaningful only in the Source objects populated |
| 5012 * by the user (e.g. when filling in a DerivedSource). |
| 5013 * Source objects supplied by the framework to the user don't have |
| 5014 * this field populated. |
4770 */ | 5015 */ |
4771 core.bool doesNotNeedSplitting; | 5016 core.bool doesNotNeedSplitting; |
4772 /** | 5017 /** |
4773 * Optionally, metadata for this source can be supplied right away, avoiding a | 5018 * Optionally, metadata for this source can be supplied right away, |
4774 * SourceGetMetadataOperation roundtrip (see SourceOperationRequest). This | 5019 * avoiding a SourceGetMetadataOperation roundtrip |
4775 * field is meaningful only in the Source objects populated by the user (e.g. | 5020 * (see SourceOperationRequest). |
4776 * when filling in a DerivedSource). Source objects supplied by the framework | 5021 * |
4777 * to the user don't have this field populated. | 5022 * This field is meaningful only in the Source objects populated |
| 5023 * by the user (e.g. when filling in a DerivedSource). |
| 5024 * Source objects supplied by the framework to the user don't have |
| 5025 * this field populated. |
4778 */ | 5026 */ |
4779 SourceMetadata metadata; | 5027 SourceMetadata metadata; |
4780 /** | 5028 /** |
4781 * The source to read from, plus its parameters. | 5029 * The source to read from, plus its parameters. |
4782 * | 5030 * |
4783 * The values for Object must be JSON objects. It can consist of `num`, | 5031 * The values for Object must be JSON objects. It can consist of `num`, |
4784 * `String`, `bool` and `null` as well as `Map` and `List` values. | 5032 * `String`, `bool` and `null` as well as `Map` and `List` values. |
4785 */ | 5033 */ |
4786 core.Map<core.String, core.Object> spec; | 5034 core.Map<core.String, core.Object> spec; |
4787 | 5035 |
(...skipping 122 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
4910 core.Map toJson() { | 5158 core.Map toJson() { |
4911 var _json = new core.Map(); | 5159 var _json = new core.Map(); |
4912 if (metadata != null) { | 5160 if (metadata != null) { |
4913 _json["metadata"] = (metadata).toJson(); | 5161 _json["metadata"] = (metadata).toJson(); |
4914 } | 5162 } |
4915 return _json; | 5163 return _json; |
4916 } | 5164 } |
4917 } | 5165 } |
4918 | 5166 |
4919 /** | 5167 /** |
4920 * Metadata about a Source useful for automatically optimizing and tuning the | 5168 * Metadata about a Source useful for automatically optimizing |
4921 * pipeline, etc. | 5169 * and tuning the pipeline, etc. |
4922 */ | 5170 */ |
4923 class SourceMetadata { | 5171 class SourceMetadata { |
4924 /** | 5172 /** |
4925 * An estimate of the total size (in bytes) of the data that would be read | 5173 * An estimate of the total size (in bytes) of the data that would be |
4926 * from this source. This estimate is in terms of external storage size, | 5174 * read from this source. This estimate is in terms of external storage |
4927 * before any decompression or other processing done by the reader. | 5175 * size, before any decompression or other processing done by the reader. |
4928 */ | 5176 */ |
4929 core.String estimatedSizeBytes; | 5177 core.String estimatedSizeBytes; |
4930 /** | 5178 /** |
4931 * Specifies that the size of this source is known to be infinite (this is a | 5179 * Specifies that the size of this source is known to be infinite |
4932 * streaming source). | 5180 * (this is a streaming source). |
4933 */ | 5181 */ |
4934 core.bool infinite; | 5182 core.bool infinite; |
4935 /** | 5183 /** |
4936 * Whether this source is known to produce key/value pairs with the (encoded) | 5184 * Whether this source is known to produce key/value pairs with |
4937 * keys in lexicographically sorted order. | 5185 * the (encoded) keys in lexicographically sorted order. |
4938 */ | 5186 */ |
4939 core.bool producesSortedKeys; | 5187 core.bool producesSortedKeys; |
4940 | 5188 |
4941 SourceMetadata(); | 5189 SourceMetadata(); |
4942 | 5190 |
4943 SourceMetadata.fromJson(core.Map _json) { | 5191 SourceMetadata.fromJson(core.Map _json) { |
4944 if (_json.containsKey("estimatedSizeBytes")) { | 5192 if (_json.containsKey("estimatedSizeBytes")) { |
4945 estimatedSizeBytes = _json["estimatedSizeBytes"]; | 5193 estimatedSizeBytes = _json["estimatedSizeBytes"]; |
4946 } | 5194 } |
4947 if (_json.containsKey("infinite")) { | 5195 if (_json.containsKey("infinite")) { |
(...skipping 13 matching lines...) Expand all Loading... |
4961 _json["infinite"] = infinite; | 5209 _json["infinite"] = infinite; |
4962 } | 5210 } |
4963 if (producesSortedKeys != null) { | 5211 if (producesSortedKeys != null) { |
4964 _json["producesSortedKeys"] = producesSortedKeys; | 5212 _json["producesSortedKeys"] = producesSortedKeys; |
4965 } | 5213 } |
4966 return _json; | 5214 return _json; |
4967 } | 5215 } |
4968 } | 5216 } |
4969 | 5217 |
4970 /** | 5218 /** |
4971 * A work item that represents the different operations that can be performed on | 5219 * A work item that represents the different operations that can be |
4972 * a user-defined Source specification. | 5220 * performed on a user-defined Source specification. |
4973 */ | 5221 */ |
4974 class SourceOperationRequest { | 5222 class SourceOperationRequest { |
4975 /** Information about a request to get metadata about a source. */ | 5223 /** Information about a request to get metadata about a source. */ |
4976 SourceGetMetadataRequest getMetadata; | 5224 SourceGetMetadataRequest getMetadata; |
4977 /** Information about a request to split a source. */ | 5225 /** Information about a request to split a source. */ |
4978 SourceSplitRequest split; | 5226 SourceSplitRequest split; |
4979 | 5227 |
4980 SourceOperationRequest(); | 5228 SourceOperationRequest(); |
4981 | 5229 |
4982 SourceOperationRequest.fromJson(core.Map _json) { | 5230 SourceOperationRequest.fromJson(core.Map _json) { |
(...skipping 12 matching lines...) Expand all Loading... |
4995 } | 5243 } |
4996 if (split != null) { | 5244 if (split != null) { |
4997 _json["split"] = (split).toJson(); | 5245 _json["split"] = (split).toJson(); |
4998 } | 5246 } |
4999 return _json; | 5247 return _json; |
5000 } | 5248 } |
5001 } | 5249 } |
5002 | 5250 |
5003 /** | 5251 /** |
5004 * The result of a SourceOperationRequest, specified in | 5252 * The result of a SourceOperationRequest, specified in |
5005 * ReportWorkItemStatusRequest.source_operation when the work item is completed. | 5253 * ReportWorkItemStatusRequest.source_operation when the work item |
| 5254 * is completed. |
5006 */ | 5255 */ |
5007 class SourceOperationResponse { | 5256 class SourceOperationResponse { |
5008 /** A response to a request to get metadata about a source. */ | 5257 /** A response to a request to get metadata about a source. */ |
5009 SourceGetMetadataResponse getMetadata; | 5258 SourceGetMetadataResponse getMetadata; |
5010 /** A response to a request to split a source. */ | 5259 /** A response to a request to split a source. */ |
5011 SourceSplitResponse split; | 5260 SourceSplitResponse split; |
5012 | 5261 |
5013 SourceOperationResponse(); | 5262 SourceOperationResponse(); |
5014 | 5263 |
5015 SourceOperationResponse.fromJson(core.Map _json) { | 5264 SourceOperationResponse.fromJson(core.Map _json) { |
(...skipping 11 matching lines...) Expand all Loading... |
5027 _json["getMetadata"] = (getMetadata).toJson(); | 5276 _json["getMetadata"] = (getMetadata).toJson(); |
5028 } | 5277 } |
5029 if (split != null) { | 5278 if (split != null) { |
5030 _json["split"] = (split).toJson(); | 5279 _json["split"] = (split).toJson(); |
5031 } | 5280 } |
5032 return _json; | 5281 return _json; |
5033 } | 5282 } |
5034 } | 5283 } |
5035 | 5284 |
5036 /** | 5285 /** |
5037 * Hints for splitting a Source into bundles (parts for parallel processing) | 5286 * Hints for splitting a Source into bundles (parts for parallel |
5038 * using SourceSplitRequest. | 5287 * processing) using SourceSplitRequest. |
5039 */ | 5288 */ |
5040 class SourceSplitOptions { | 5289 class SourceSplitOptions { |
5041 /** | 5290 /** |
5042 * The source should be split into a set of bundles where the estimated size | 5291 * The source should be split into a set of bundles where the estimated size |
5043 * of each is approximately this many bytes. | 5292 * of each is approximately this many bytes. |
5044 */ | 5293 */ |
5045 core.String desiredBundleSizeBytes; | 5294 core.String desiredBundleSizeBytes; |
5046 /** DEPRECATED in favor of desired_bundle_size_bytes. */ | 5295 /** DEPRECATED in favor of desired_bundle_size_bytes. */ |
5047 core.String desiredShardSizeBytes; | 5296 core.String desiredShardSizeBytes; |
5048 | 5297 |
(...skipping 14 matching lines...) Expand all Loading... |
5063 _json["desiredBundleSizeBytes"] = desiredBundleSizeBytes; | 5312 _json["desiredBundleSizeBytes"] = desiredBundleSizeBytes; |
5064 } | 5313 } |
5065 if (desiredShardSizeBytes != null) { | 5314 if (desiredShardSizeBytes != null) { |
5066 _json["desiredShardSizeBytes"] = desiredShardSizeBytes; | 5315 _json["desiredShardSizeBytes"] = desiredShardSizeBytes; |
5067 } | 5316 } |
5068 return _json; | 5317 return _json; |
5069 } | 5318 } |
5070 } | 5319 } |
5071 | 5320 |
5072 /** | 5321 /** |
5073 * Represents the operation to split a high-level Source specification into | 5322 * Represents the operation to split a high-level Source specification |
5074 * bundles (parts for parallel processing). At a high level, splitting of a | 5323 * into bundles (parts for parallel processing). |
5075 * source into bundles happens as follows: SourceSplitRequest is applied to the | 5324 * |
5076 * source. If it returns SOURCE_SPLIT_OUTCOME_USE_CURRENT, no further splitting | 5325 * At a high level, splitting of a source into bundles happens as follows: |
5077 * happens and the source is used "as is". Otherwise, splitting is applied | 5326 * SourceSplitRequest is applied to the source. If it returns |
5078 * recursively to each produced DerivedSource. As an optimization, for any | 5327 * SOURCE_SPLIT_OUTCOME_USE_CURRENT, no further splitting happens and the source |
5079 * Source, if its does_not_need_splitting is true, the framework assumes that | 5328 * is used "as is". Otherwise, splitting is applied recursively to each |
5080 * splitting this source would return SOURCE_SPLIT_OUTCOME_USE_CURRENT, and | 5329 * produced DerivedSource. |
5081 * doesn't initiate a SourceSplitRequest. This applies both to the initial | 5330 * |
5082 * source being split and to bundles produced from it. | 5331 * As an optimization, for any Source, if its does_not_need_splitting is |
| 5332 * true, the framework assumes that splitting this source would return |
| 5333 * SOURCE_SPLIT_OUTCOME_USE_CURRENT, and doesn't initiate a SourceSplitRequest. |
| 5334 * This applies both to the initial source being split and to bundles |
| 5335 * produced from it. |
5083 */ | 5336 */ |
5084 class SourceSplitRequest { | 5337 class SourceSplitRequest { |
5085 /** Hints for tuning the splitting process. */ | 5338 /** Hints for tuning the splitting process. */ |
5086 SourceSplitOptions options; | 5339 SourceSplitOptions options; |
5087 /** Specification of the source to be split. */ | 5340 /** Specification of the source to be split. */ |
5088 Source source; | 5341 Source source; |
5089 | 5342 |
5090 SourceSplitRequest(); | 5343 SourceSplitRequest(); |
5091 | 5344 |
5092 SourceSplitRequest.fromJson(core.Map _json) { | 5345 SourceSplitRequest.fromJson(core.Map _json) { |
(...skipping 13 matching lines...) Expand all Loading... |
5106 if (source != null) { | 5359 if (source != null) { |
5107 _json["source"] = (source).toJson(); | 5360 _json["source"] = (source).toJson(); |
5108 } | 5361 } |
5109 return _json; | 5362 return _json; |
5110 } | 5363 } |
5111 } | 5364 } |
5112 | 5365 |
5113 /** The response to a SourceSplitRequest. */ | 5366 /** The response to a SourceSplitRequest. */ |
5114 class SourceSplitResponse { | 5367 class SourceSplitResponse { |
5115 /** | 5368 /** |
5116 * If outcome is SPLITTING_HAPPENED, then this is a list of bundles into which | 5369 * If outcome is SPLITTING_HAPPENED, then this is a list of bundles |
5117 * the source was split. Otherwise this field is ignored. This list can be | 5370 * into which the source was split. Otherwise this field is ignored. |
5118 * empty, which means the source represents an empty input. | 5371 * This list can be empty, which means the source represents an empty input. |
5119 */ | 5372 */ |
5120 core.List<DerivedSource> bundles; | 5373 core.List<DerivedSource> bundles; |
5121 /** | 5374 /** |
5122 * Indicates whether splitting happened and produced a list of bundles. If | 5375 * Indicates whether splitting happened and produced a list of bundles. |
5123 * this is USE_CURRENT_SOURCE_AS_IS, the current source should be processed | 5376 * If this is USE_CURRENT_SOURCE_AS_IS, the current source should |
5124 * "as is" without splitting. "bundles" is ignored in this case. If this is | 5377 * be processed "as is" without splitting. "bundles" is ignored in this case. |
5125 * SPLITTING_HAPPENED, then "bundles" contains a list of bundles into which | 5378 * If this is SPLITTING_HAPPENED, then "bundles" contains a list of |
5126 * the source was split. | 5379 * bundles into which the source was split. |
5127 * Possible string values are: | 5380 * Possible string values are: |
5128 * - "SOURCE_SPLIT_OUTCOME_UNKNOWN" : A SOURCE_SPLIT_OUTCOME_UNKNOWN. | 5381 * - "SOURCE_SPLIT_OUTCOME_UNKNOWN" : The source split outcome is unknown, or |
5129 * - "SOURCE_SPLIT_OUTCOME_USE_CURRENT" : A SOURCE_SPLIT_OUTCOME_USE_CURRENT. | 5382 * unspecified. |
5130 * - "SOURCE_SPLIT_OUTCOME_SPLITTING_HAPPENED" : A | 5383 * - "SOURCE_SPLIT_OUTCOME_USE_CURRENT" : The current source should be |
5131 * SOURCE_SPLIT_OUTCOME_SPLITTING_HAPPENED. | 5384 * processed "as is" without splitting. |
| 5385 * - "SOURCE_SPLIT_OUTCOME_SPLITTING_HAPPENED" : Splitting produced a list of |
| 5386 * bundles. |
5132 */ | 5387 */ |
5133 core.String outcome; | 5388 core.String outcome; |
5134 /** DEPRECATED in favor of bundles. */ | 5389 /** DEPRECATED in favor of bundles. */ |
5135 core.List<SourceSplitShard> shards; | 5390 core.List<SourceSplitShard> shards; |
5136 | 5391 |
5137 SourceSplitResponse(); | 5392 SourceSplitResponse(); |
5138 | 5393 |
5139 SourceSplitResponse.fromJson(core.Map _json) { | 5394 SourceSplitResponse.fromJson(core.Map _json) { |
5140 if (_json.containsKey("bundles")) { | 5395 if (_json.containsKey("bundles")) { |
5141 bundles = _json["bundles"].map((value) => new DerivedSource.fromJson(value
)).toList(); | 5396 bundles = _json["bundles"].map((value) => new DerivedSource.fromJson(value
)).toList(); |
(...skipping 19 matching lines...) Expand all Loading... |
5161 } | 5416 } |
5162 return _json; | 5417 return _json; |
5163 } | 5418 } |
5164 } | 5419 } |
5165 | 5420 |
5166 /** DEPRECATED in favor of DerivedSource. */ | 5421 /** DEPRECATED in favor of DerivedSource. */ |
5167 class SourceSplitShard { | 5422 class SourceSplitShard { |
5168 /** | 5423 /** |
5169 * DEPRECATED | 5424 * DEPRECATED |
5170 * Possible string values are: | 5425 * Possible string values are: |
5171 * - "SOURCE_DERIVATION_MODE_UNKNOWN" : A SOURCE_DERIVATION_MODE_UNKNOWN. | 5426 * - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or |
5172 * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : A | 5427 * unspecified. |
5173 * SOURCE_DERIVATION_MODE_INDEPENDENT. | 5428 * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent |
5174 * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : A | 5429 * Source with no base. |
5175 * SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT. | 5430 * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on the |
5176 * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : A | 5431 * Source being split. |
5177 * SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT. | 5432 * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on |
| 5433 * the base of the Source being split. |
5178 */ | 5434 */ |
5179 core.String derivationMode; | 5435 core.String derivationMode; |
5180 /** DEPRECATED */ | 5436 /** DEPRECATED */ |
5181 Source source; | 5437 Source source; |
5182 | 5438 |
5183 SourceSplitShard(); | 5439 SourceSplitShard(); |
5184 | 5440 |
5185 SourceSplitShard.fromJson(core.Map _json) { | 5441 SourceSplitShard.fromJson(core.Map _json) { |
5186 if (_json.containsKey("derivationMode")) { | 5442 if (_json.containsKey("derivationMode")) { |
5187 derivationMode = _json["derivationMode"]; | 5443 derivationMode = _json["derivationMode"]; |
(...skipping 73 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5261 } | 5517 } |
5262 if (stateFamily != null) { | 5518 if (stateFamily != null) { |
5263 _json["stateFamily"] = stateFamily; | 5519 _json["stateFamily"] = stateFamily; |
5264 } | 5520 } |
5265 return _json; | 5521 return _json; |
5266 } | 5522 } |
5267 } | 5523 } |
5268 | 5524 |
5269 /** | 5525 /** |
5270 * The `Status` type defines a logical error model that is suitable for | 5526 * The `Status` type defines a logical error model that is suitable for |
5271 * different programming environments, including REST APIs and RPC APIs. It is | 5527 * different |
5272 * used by [gRPC](https://github.com/grpc). The error model is designed to be: - | 5528 * programming environments, including REST APIs and RPC APIs. It is used by |
5273 * Simple to use and understand for most users - Flexible enough to meet | 5529 * [gRPC](https://github.com/grpc). The error model is designed to be: |
5274 * unexpected needs # Overview The `Status` message contains three pieces of | 5530 * |
5275 * data: error code, error message, and error details. The error code should be | 5531 * - Simple to use and understand for most users |
5276 * an enum value of google.rpc.Code, but it may accept additional error codes if | 5532 * - Flexible enough to meet unexpected needs |
5277 * needed. The error message should be a developer-facing English message that | 5533 * |
5278 * helps developers *understand* and *resolve* the error. If a localized | 5534 * # Overview |
5279 * user-facing error message is needed, put the localized message in the error | 5535 * |
5280 * details or localize it in the client. The optional error details may contain | 5536 * The `Status` message contains three pieces of data: error code, error |
5281 * arbitrary information about the error. There is a predefined set of error | 5537 * message, |
5282 * detail types in the package `google.rpc` which can be used for common error | 5538 * and error details. The error code should be an enum value of |
5283 * conditions. # Language mapping The `Status` message is the logical | 5539 * google.rpc.Code, but it may accept additional error codes if needed. The |
5284 * representation of the error model, but it is not necessarily the actual wire | 5540 * error message should be a developer-facing English message that helps |
5285 * format. When the `Status` message is exposed in different client libraries | 5541 * developers *understand* and *resolve* the error. If a localized user-facing |
5286 * and different wire protocols, it can be mapped differently. For example, it | 5542 * error message is needed, put the localized message in the error details or |
5287 * will likely be mapped to some exceptions in Java, but more likely mapped to | 5543 * localize it in the client. The optional error details may contain arbitrary |
5288 * some error codes in C. # Other uses The error model and the `Status` message | 5544 * information about the error. There is a predefined set of error detail types |
5289 * can be used in a variety of environments, either with or without APIs, to | 5545 * in the package `google.rpc` which can be used for common error conditions. |
5290 * provide a consistent developer experience across different environments. | 5546 * |
5291 * Example uses of this error model include: - Partial errors. If a service | 5547 * # Language mapping |
5292 * needs to return partial errors to the client, it may embed the `Status` in | 5548 * |
5293 * the normal response to indicate the partial errors. - Workflow errors. A | 5549 * The `Status` message is the logical representation of the error model, but it |
5294 * typical workflow has multiple steps. Each step may have a `Status` message | 5550 * is not necessarily the actual wire format. When the `Status` message is |
5295 * for error reporting purpose. - Batch operations. If a client uses batch | 5551 * exposed in different client libraries and different wire protocols, it can be |
5296 * request and batch response, the `Status` message should be used directly | 5552 * mapped differently. For example, it will likely be mapped to some exceptions |
5297 * inside batch response, one for each error sub-response. - Asynchronous | 5553 * in Java, but more likely mapped to some error codes in C. |
5298 * operations. If an API call embeds asynchronous operation results in its | 5554 * |
5299 * response, the status of those operations should be represented directly using | 5555 * # Other uses |
5300 * the `Status` message. - Logging. If some API errors are stored in logs, the | 5556 * |
5301 * message `Status` could be used directly after any stripping needed for | 5557 * The error model and the `Status` message can be used in a variety of |
5302 * security/privacy reasons. | 5558 * environments, either with or without APIs, to provide a |
| 5559 * consistent developer experience across different environments. |
| 5560 * |
| 5561 * Example uses of this error model include: |
| 5562 * |
| 5563 * - Partial errors. If a service needs to return partial errors to the client, |
| 5564 * it may embed the `Status` in the normal response to indicate the partial |
| 5565 * errors. |
| 5566 * |
| 5567 * - Workflow errors. A typical workflow has multiple steps. Each step may |
| 5568 * have a `Status` message for error reporting purpose. |
| 5569 * |
| 5570 * - Batch operations. If a client uses batch request and batch response, the |
| 5571 * `Status` message should be used directly inside batch response, one for |
| 5572 * each error sub-response. |
| 5573 * |
| 5574 * - Asynchronous operations. If an API call embeds asynchronous operation |
| 5575 * results in its response, the status of those operations should be |
| 5576 * represented directly using the `Status` message. |
| 5577 * |
| 5578 * - Logging. If some API errors are stored in logs, the message `Status` could |
| 5579 * be used directly after any stripping needed for security/privacy reasons. |
5303 */ | 5580 */ |
5304 class Status { | 5581 class Status { |
5305 /** The status code, which should be an enum value of google.rpc.Code. */ | 5582 /** The status code, which should be an enum value of google.rpc.Code. */ |
5306 core.int code; | 5583 core.int code; |
5307 /** | 5584 /** |
5308 * A list of messages that carry the error details. There will be a common set | 5585 * A list of messages that carry the error details. There will be a |
5309 * of message types for APIs to use. | 5586 * common set of message types for APIs to use. |
5310 * | 5587 * |
5311 * The values for Object must be JSON objects. It can consist of `num`, | 5588 * The values for Object must be JSON objects. It can consist of `num`, |
5312 * `String`, `bool` and `null` as well as `Map` and `List` values. | 5589 * `String`, `bool` and `null` as well as `Map` and `List` values. |
5313 */ | 5590 */ |
5314 core.List<core.Map<core.String, core.Object>> details; | 5591 core.List<core.Map<core.String, core.Object>> details; |
5315 /** | 5592 /** |
5316 * A developer-facing error message, which should be in English. Any | 5593 * A developer-facing error message, which should be in English. Any |
5317 * user-facing error message should be localized and sent in the | 5594 * user-facing error message should be localized and sent in the |
5318 * google.rpc.Status.details field, or localized by the client. | 5595 * google.rpc.Status.details field, or localized by the client. |
5319 */ | 5596 */ |
(...skipping 22 matching lines...) Expand all Loading... |
5342 _json["details"] = details; | 5619 _json["details"] = details; |
5343 } | 5620 } |
5344 if (message != null) { | 5621 if (message != null) { |
5345 _json["message"] = message; | 5622 _json["message"] = message; |
5346 } | 5623 } |
5347 return _json; | 5624 return _json; |
5348 } | 5625 } |
5349 } | 5626 } |
5350 | 5627 |
5351 /** | 5628 /** |
5352 * Defines a particular step within a Cloud Dataflow job. A job consists of | 5629 * Defines a particular step within a Cloud Dataflow job. |
5353 * multiple steps, each of which performs some specific operation as part of the | 5630 * |
5354 * overall job. Data is typically passed from one step to another as part of the | 5631 * A job consists of multiple steps, each of which performs some |
5355 * job. Here's an example of a sequence of steps which together implement a | 5632 * specific operation as part of the overall job. Data is typically |
5356 * Map-Reduce job: * Read a collection of data from some source, parsing the | 5633 * passed from one step to another as part of the job. |
5357 * collection's elements. * Validate the elements. * Apply a user-defined | 5634 * |
5358 * function to map each element to some value and extract an element-specific | 5635 * Here's an example of a sequence of steps which together implement a |
5359 * key value. * Group elements with the same key into a single element with that | 5636 * Map-Reduce job: |
5360 * key, transforming a multiply-keyed collection into a uniquely-keyed | 5637 * |
5361 * collection. * Write the elements out to some data sink. Note that the Cloud | 5638 * * Read a collection of data from some source, parsing the |
5362 * Dataflow service may be used to run many different types of jobs, not just | 5639 * collection's elements. |
5363 * Map-Reduce. | 5640 * |
| 5641 * * Validate the elements. |
| 5642 * |
| 5643 * * Apply a user-defined function to map each element to some value |
| 5644 * and extract an element-specific key value. |
| 5645 * |
| 5646 * * Group elements with the same key into a single element with |
| 5647 * that key, transforming a multiply-keyed collection into a |
| 5648 * uniquely-keyed collection. |
| 5649 * |
| 5650 * * Write the elements out to some data sink. |
| 5651 * |
| 5652 * Note that the Cloud Dataflow service may be used to run many different |
| 5653 * types of jobs, not just Map-Reduce. |
5364 */ | 5654 */ |
5365 class Step { | 5655 class Step { |
5366 /** The kind of step in the Cloud Dataflow job. */ | 5656 /** The kind of step in the Cloud Dataflow job. */ |
5367 core.String kind; | 5657 core.String kind; |
5368 /** | 5658 /** |
5369 * The name that identifies the step. This must be unique for each step with | 5659 * The name that identifies the step. This must be unique for each |
5370 * respect to all other steps in the Cloud Dataflow job. | 5660 * step with respect to all other steps in the Cloud Dataflow job. |
5371 */ | 5661 */ |
5372 core.String name; | 5662 core.String name; |
5373 /** | 5663 /** |
5374 * Named properties associated with the step. Each kind of predefined step has | 5664 * Named properties associated with the step. Each kind of |
5375 * its own required set of properties. | 5665 * predefined step has its own required set of properties. |
5376 * | 5666 * |
5377 * The values for Object must be JSON objects. It can consist of `num`, | 5667 * The values for Object must be JSON objects. It can consist of `num`, |
5378 * `String`, `bool` and `null` as well as `Map` and `List` values. | 5668 * `String`, `bool` and `null` as well as `Map` and `List` values. |
5379 */ | 5669 */ |
5380 core.Map<core.String, core.Object> properties; | 5670 core.Map<core.String, core.Object> properties; |
5381 | 5671 |
5382 Step(); | 5672 Step(); |
5383 | 5673 |
5384 Step.fromJson(core.Map _json) { | 5674 Step.fromJson(core.Map _json) { |
5385 if (_json.containsKey("kind")) { | 5675 if (_json.containsKey("kind")) { |
(...skipping 16 matching lines...) Expand all Loading... |
5402 _json["name"] = name; | 5692 _json["name"] = name; |
5403 } | 5693 } |
5404 if (properties != null) { | 5694 if (properties != null) { |
5405 _json["properties"] = properties; | 5695 _json["properties"] = properties; |
5406 } | 5696 } |
5407 return _json; | 5697 return _json; |
5408 } | 5698 } |
5409 } | 5699 } |
5410 | 5700 |
5411 /** | 5701 /** |
5412 * Describes a stream of data, either as input to be processed or as output of a | 5702 * Describes a stream of data, either as input to be processed or as |
5413 * streaming Dataflow job. | 5703 * output of a streaming Dataflow job. |
5414 */ | 5704 */ |
5415 class StreamLocation { | 5705 class StreamLocation { |
5416 /** The stream is a custom source. */ | 5706 /** The stream is a custom source. */ |
5417 CustomSourceLocation customSourceLocation; | 5707 CustomSourceLocation customSourceLocation; |
5418 /** The stream is a pubsub stream. */ | 5708 /** The stream is a pubsub stream. */ |
5419 PubsubLocation pubsubLocation; | 5709 PubsubLocation pubsubLocation; |
5420 /** The stream is a streaming side input. */ | 5710 /** The stream is a streaming side input. */ |
5421 StreamingSideInputLocation sideInputLocation; | 5711 StreamingSideInputLocation sideInputLocation; |
5422 /** | 5712 /** |
5423 * The stream is part of another computation within the current streaming | 5713 * The stream is part of another computation within the current |
5424 * Dataflow job. | 5714 * streaming Dataflow job. |
5425 */ | 5715 */ |
5426 StreamingStageLocation streamingStageLocation; | 5716 StreamingStageLocation streamingStageLocation; |
5427 | 5717 |
5428 StreamLocation(); | 5718 StreamLocation(); |
5429 | 5719 |
5430 StreamLocation.fromJson(core.Map _json) { | 5720 StreamLocation.fromJson(core.Map _json) { |
5431 if (_json.containsKey("customSourceLocation")) { | 5721 if (_json.containsKey("customSourceLocation")) { |
5432 customSourceLocation = new CustomSourceLocation.fromJson(_json["customSour
ceLocation"]); | 5722 customSourceLocation = new CustomSourceLocation.fromJson(_json["customSour
ceLocation"]); |
5433 } | 5723 } |
5434 if (_json.containsKey("pubsubLocation")) { | 5724 if (_json.containsKey("pubsubLocation")) { |
(...skipping 109 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5544 * streaming computation ranges. | 5834 * streaming computation ranges. |
5545 */ | 5835 */ |
5546 class StreamingComputationTask { | 5836 class StreamingComputationTask { |
5547 /** Contains ranges of a streaming computation this task should apply to. */ | 5837 /** Contains ranges of a streaming computation this task should apply to. */ |
5548 core.List<StreamingComputationRanges> computationRanges; | 5838 core.List<StreamingComputationRanges> computationRanges; |
5549 /** Describes the set of data disks this task should apply to. */ | 5839 /** Describes the set of data disks this task should apply to. */ |
5550 core.List<MountedDataDisk> dataDisks; | 5840 core.List<MountedDataDisk> dataDisks; |
5551 /** | 5841 /** |
5552 * A type of streaming computation task. | 5842 * A type of streaming computation task. |
5553 * Possible string values are: | 5843 * Possible string values are: |
5554 * - "STREAMING_COMPUTATION_TASK_UNKNOWN" : A | 5844 * - "STREAMING_COMPUTATION_TASK_UNKNOWN" : The streaming computation task is |
5555 * STREAMING_COMPUTATION_TASK_UNKNOWN. | 5845 * unknown, or unspecified. |
5556 * - "STREAMING_COMPUTATION_TASK_STOP" : A STREAMING_COMPUTATION_TASK_STOP. | 5846 * - "STREAMING_COMPUTATION_TASK_STOP" : Stop processing specified streaming |
5557 * - "STREAMING_COMPUTATION_TASK_START" : A STREAMING_COMPUTATION_TASK_START. | 5847 * computation range(s). |
| 5848 * - "STREAMING_COMPUTATION_TASK_START" : Start processing specified streaming |
| 5849 * computation range(s). |
5558 */ | 5850 */ |
5559 core.String taskType; | 5851 core.String taskType; |
5560 | 5852 |
5561 StreamingComputationTask(); | 5853 StreamingComputationTask(); |
5562 | 5854 |
5563 StreamingComputationTask.fromJson(core.Map _json) { | 5855 StreamingComputationTask.fromJson(core.Map _json) { |
5564 if (_json.containsKey("computationRanges")) { | 5856 if (_json.containsKey("computationRanges")) { |
5565 computationRanges = _json["computationRanges"].map((value) => new Streamin
gComputationRanges.fromJson(value)).toList(); | 5857 computationRanges = _json["computationRanges"].map((value) => new Streamin
gComputationRanges.fromJson(value)).toList(); |
5566 } | 5858 } |
5567 if (_json.containsKey("dataDisks")) { | 5859 if (_json.containsKey("dataDisks")) { |
(...skipping 49 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5617 } | 5909 } |
5618 return _json; | 5910 return _json; |
5619 } | 5911 } |
5620 } | 5912 } |
5621 | 5913 |
5622 /** A task which initializes part of a streaming Dataflow job. */ | 5914 /** A task which initializes part of a streaming Dataflow job. */ |
5623 class StreamingSetupTask { | 5915 class StreamingSetupTask { |
5624 /** The user has requested drain. */ | 5916 /** The user has requested drain. */ |
5625 core.bool drain; | 5917 core.bool drain; |
5626 /** | 5918 /** |
5627 * The TCP port on which the worker should listen for messages from other | 5919 * The TCP port on which the worker should listen for messages from |
5628 * streaming computation workers. | 5920 * other streaming computation workers. |
5629 */ | 5921 */ |
5630 core.int receiveWorkPort; | 5922 core.int receiveWorkPort; |
5631 /** The global topology of the streaming Dataflow job. */ | 5923 /** The global topology of the streaming Dataflow job. */ |
5632 TopologyConfig streamingComputationTopology; | 5924 TopologyConfig streamingComputationTopology; |
5633 /** | 5925 /** |
5634 * The TCP port used by the worker to communicate with the Dataflow worker | 5926 * The TCP port used by the worker to communicate with the Dataflow |
5635 * harness. | 5927 * worker harness. |
5636 */ | 5928 */ |
5637 core.int workerHarnessPort; | 5929 core.int workerHarnessPort; |
5638 | 5930 |
5639 StreamingSetupTask(); | 5931 StreamingSetupTask(); |
5640 | 5932 |
5641 StreamingSetupTask.fromJson(core.Map _json) { | 5933 StreamingSetupTask.fromJson(core.Map _json) { |
5642 if (_json.containsKey("drain")) { | 5934 if (_json.containsKey("drain")) { |
5643 drain = _json["drain"]; | 5935 drain = _json["drain"]; |
5644 } | 5936 } |
5645 if (_json.containsKey("receiveWorkPort")) { | 5937 if (_json.containsKey("receiveWorkPort")) { |
(...skipping 51 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5697 _json["stateFamily"] = stateFamily; | 5989 _json["stateFamily"] = stateFamily; |
5698 } | 5990 } |
5699 if (tag != null) { | 5991 if (tag != null) { |
5700 _json["tag"] = tag; | 5992 _json["tag"] = tag; |
5701 } | 5993 } |
5702 return _json; | 5994 return _json; |
5703 } | 5995 } |
5704 } | 5996 } |
5705 | 5997 |
5706 /** | 5998 /** |
5707 * Identifies the location of a streaming computation stage, for stage-to-stage | 5999 * Identifies the location of a streaming computation stage, for |
5708 * communication. | 6000 * stage-to-stage communication. |
5709 */ | 6001 */ |
5710 class StreamingStageLocation { | 6002 class StreamingStageLocation { |
5711 /** Identifies the particular stream within the streaming Dataflow job. */ | 6003 /** |
| 6004 * Identifies the particular stream within the streaming Dataflow |
| 6005 * job. |
| 6006 */ |
5712 core.String streamId; | 6007 core.String streamId; |
5713 | 6008 |
5714 StreamingStageLocation(); | 6009 StreamingStageLocation(); |
5715 | 6010 |
5716 StreamingStageLocation.fromJson(core.Map _json) { | 6011 StreamingStageLocation.fromJson(core.Map _json) { |
5717 if (_json.containsKey("streamId")) { | 6012 if (_json.containsKey("streamId")) { |
5718 streamId = _json["streamId"]; | 6013 streamId = _json["streamId"]; |
5719 } | 6014 } |
5720 } | 6015 } |
5721 | 6016 |
(...skipping 29 matching lines...) Expand all Loading... |
5751 } | 6046 } |
5752 | 6047 |
5753 /** Taskrunner configuration settings. */ | 6048 /** Taskrunner configuration settings. */ |
5754 class TaskRunnerSettings { | 6049 class TaskRunnerSettings { |
5755 /** Whether to also send taskrunner log info to stderr. */ | 6050 /** Whether to also send taskrunner log info to stderr. */ |
5756 core.bool alsologtostderr; | 6051 core.bool alsologtostderr; |
5757 /** The location on the worker for task-specific subdirectories. */ | 6052 /** The location on the worker for task-specific subdirectories. */ |
5758 core.String baseTaskDir; | 6053 core.String baseTaskDir; |
5759 /** | 6054 /** |
5760 * The base URL for the taskrunner to use when accessing Google Cloud APIs. | 6055 * The base URL for the taskrunner to use when accessing Google Cloud APIs. |
5761 * When workers access Google Cloud APIs, they logically do so via relative | 6056 * |
5762 * URLs. If this field is specified, it supplies the base URL to use for | 6057 * When workers access Google Cloud APIs, they logically do so via |
5763 * resolving these relative URLs. The normative algorithm used is defined by | 6058 * relative URLs. If this field is specified, it supplies the base |
5764 * RFC 1808, "Relative Uniform Resource Locators". If not specified, the | 6059 * URL to use for resolving these relative URLs. The normative |
5765 * default value is "http://www.googleapis.com/" | 6060 * algorithm used is defined by RFC 1808, "Relative Uniform Resource |
| 6061 * Locators". |
| 6062 * |
| 6063 * If not specified, the default value is "http://www.googleapis.com/" |
5766 */ | 6064 */ |
5767 core.String baseUrl; | 6065 core.String baseUrl; |
5768 /** The file to store preprocessing commands in. */ | 6066 /** The file to store preprocessing commands in. */ |
5769 core.String commandlinesFileName; | 6067 core.String commandlinesFileName; |
5770 /** Whether to continue taskrunner if an exception is hit. */ | 6068 /** Whether to continue taskrunner if an exception is hit. */ |
5771 core.bool continueOnException; | 6069 core.bool continueOnException; |
5772 /** The API version of endpoint, e.g. "v1b3" */ | 6070 /** The API version of endpoint, e.g. "v1b3" */ |
5773 core.String dataflowApiVersion; | 6071 core.String dataflowApiVersion; |
5774 /** The command to launch the worker harness. */ | 6072 /** The command to launch the worker harness. */ |
5775 core.String harnessCommand; | 6073 core.String harnessCommand; |
5776 /** The suggested backend language. */ | 6074 /** The suggested backend language. */ |
5777 core.String languageHint; | 6075 core.String languageHint; |
5778 /** The directory on the VM to store logs. */ | 6076 /** The directory on the VM to store logs. */ |
5779 core.String logDir; | 6077 core.String logDir; |
5780 /** | 6078 /** |
5781 * Whether to send taskrunner log info to Google Compute Engine VM serial | 6079 * Whether to send taskrunner log info to Google Compute Engine VM serial |
5782 * console. | 6080 * console. |
5783 */ | 6081 */ |
5784 core.bool logToSerialconsole; | 6082 core.bool logToSerialconsole; |
5785 /** | 6083 /** |
5786 * Indicates where to put logs. If this is not specified, the logs will not be | 6084 * Indicates where to put logs. If this is not specified, the logs |
5787 * uploaded. The supported resource type is: Google Cloud Storage: | 6085 * will not be uploaded. |
5788 * storage.googleapis.com/{bucket}/{object} | 6086 * |
5789 * bucket.storage.googleapis.com/{object} | 6087 * The supported resource type is: |
| 6088 * |
| 6089 * Google Cloud Storage: |
| 6090 * storage.googleapis.com/{bucket}/{object} |
| 6091 * bucket.storage.googleapis.com/{object} |
5790 */ | 6092 */ |
5791 core.String logUploadLocation; | 6093 core.String logUploadLocation; |
5792 /** | 6094 /** |
5793 * The OAuth2 scopes to be requested by the taskrunner in order to access the | 6095 * The OAuth2 scopes to be requested by the taskrunner in order to |
5794 * Cloud Dataflow API. | 6096 * access the Cloud Dataflow API. |
5795 */ | 6097 */ |
5796 core.List<core.String> oauthScopes; | 6098 core.List<core.String> oauthScopes; |
5797 /** The settings to pass to the parallel worker harness. */ | 6099 /** The settings to pass to the parallel worker harness. */ |
5798 WorkerSettings parallelWorkerSettings; | 6100 WorkerSettings parallelWorkerSettings; |
5799 /** The streaming worker main class name. */ | 6101 /** The streaming worker main class name. */ |
5800 core.String streamingWorkerMainClass; | 6102 core.String streamingWorkerMainClass; |
5801 /** | 6103 /** |
5802 * The UNIX group ID on the worker VM to use for tasks launched by taskrunner; | 6104 * The UNIX group ID on the worker VM to use for tasks launched by |
5803 * e.g. "wheel". | 6105 * taskrunner; e.g. "wheel". |
5804 */ | 6106 */ |
5805 core.String taskGroup; | 6107 core.String taskGroup; |
5806 /** | 6108 /** |
5807 * The UNIX user ID on the worker VM to use for tasks launched by taskrunner; | 6109 * The UNIX user ID on the worker VM to use for tasks launched by |
5808 * e.g. "root". | 6110 * taskrunner; e.g. "root". |
5809 */ | 6111 */ |
5810 core.String taskUser; | 6112 core.String taskUser; |
5811 /** | 6113 /** |
5812 * The prefix of the resources the taskrunner should use for temporary | 6114 * The prefix of the resources the taskrunner should use for |
5813 * storage. The supported resource type is: Google Cloud Storage: | 6115 * temporary storage. |
5814 * storage.googleapis.com/{bucket}/{object} | 6116 * |
5815 * bucket.storage.googleapis.com/{object} | 6117 * The supported resource type is: |
| 6118 * |
| 6119 * Google Cloud Storage: |
| 6120 * storage.googleapis.com/{bucket}/{object} |
| 6121 * bucket.storage.googleapis.com/{object} |
5816 */ | 6122 */ |
5817 core.String tempStoragePrefix; | 6123 core.String tempStoragePrefix; |
5818 /** The ID string of the VM. */ | 6124 /** The ID string of the VM. */ |
5819 core.String vmId; | 6125 core.String vmId; |
5820 /** The file to store the workflow in. */ | 6126 /** The file to store the workflow in. */ |
5821 core.String workflowFileName; | 6127 core.String workflowFileName; |
5822 | 6128 |
5823 TaskRunnerSettings(); | 6129 TaskRunnerSettings(); |
5824 | 6130 |
5825 TaskRunnerSettings.fromJson(core.Map _json) { | 6131 TaskRunnerSettings.fromJson(core.Map _json) { |
(...skipping 113 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5939 _json["vmId"] = vmId; | 6245 _json["vmId"] = vmId; |
5940 } | 6246 } |
5941 if (workflowFileName != null) { | 6247 if (workflowFileName != null) { |
5942 _json["workflowFileName"] = workflowFileName; | 6248 _json["workflowFileName"] = workflowFileName; |
5943 } | 6249 } |
5944 return _json; | 6250 return _json; |
5945 } | 6251 } |
5946 } | 6252 } |
5947 | 6253 |
5948 /** | 6254 /** |
5949 * Global topology of the streaming Dataflow job, including all computations and | 6255 * Global topology of the streaming Dataflow job, including all |
5950 * their sharded locations. | 6256 * computations and their sharded locations. |
5951 */ | 6257 */ |
5952 class TopologyConfig { | 6258 class TopologyConfig { |
5953 /** The computations associated with a streaming Dataflow job. */ | 6259 /** The computations associated with a streaming Dataflow job. */ |
5954 core.List<ComputationTopology> computations; | 6260 core.List<ComputationTopology> computations; |
5955 /** The disks assigned to a streaming Dataflow job. */ | 6261 /** The disks assigned to a streaming Dataflow job. */ |
5956 core.List<DataDiskAssignment> dataDiskAssignments; | 6262 core.List<DataDiskAssignment> dataDiskAssignments; |
5957 /** The size (in bits) of keys that will be assigned to source messages. */ | 6263 /** The size (in bits) of keys that will be assigned to source messages. */ |
5958 core.int forwardingKeyBits; | 6264 core.int forwardingKeyBits; |
5959 /** Version number for persistent state. */ | 6265 /** Version number for persistent state. */ |
5960 core.int persistentStateVersion; | 6266 core.int persistentStateVersion; |
(...skipping 35 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
5996 _json["persistentStateVersion"] = persistentStateVersion; | 6302 _json["persistentStateVersion"] = persistentStateVersion; |
5997 } | 6303 } |
5998 if (userStageToComputationNameMap != null) { | 6304 if (userStageToComputationNameMap != null) { |
5999 _json["userStageToComputationNameMap"] = userStageToComputationNameMap; | 6305 _json["userStageToComputationNameMap"] = userStageToComputationNameMap; |
6000 } | 6306 } |
6001 return _json; | 6307 return _json; |
6002 } | 6308 } |
6003 } | 6309 } |
6004 | 6310 |
6005 /** | 6311 /** |
6006 * WorkItem represents basic information about a WorkItem to be executed in the | 6312 * WorkItem represents basic information about a WorkItem to be executed |
6007 * cloud. | 6313 * in the cloud. |
6008 */ | 6314 */ |
6009 class WorkItem { | 6315 class WorkItem { |
6010 /** Work item-specific configuration as an opaque blob. */ | 6316 /** Work item-specific configuration as an opaque blob. */ |
6011 core.String configuration; | 6317 core.String configuration; |
6012 /** Identifies this WorkItem. */ | 6318 /** Identifies this WorkItem. */ |
6013 core.String id; | 6319 core.String id; |
6014 /** The initial index to use when reporting the status of the WorkItem. */ | 6320 /** The initial index to use when reporting the status of the WorkItem. */ |
6015 core.String initialReportIndex; | 6321 core.String initialReportIndex; |
6016 /** Identifies the workflow job this WorkItem belongs to. */ | 6322 /** Identifies the workflow job this WorkItem belongs to. */ |
6017 core.String jobId; | 6323 core.String jobId; |
6018 /** Time when the lease on this Work will expire. */ | 6324 /** Time when the lease on this Work will expire. */ |
6019 core.String leaseExpireTime; | 6325 core.String leaseExpireTime; |
6020 /** Additional information for MapTask WorkItems. */ | 6326 /** Additional information for MapTask WorkItems. */ |
6021 MapTask mapTask; | 6327 MapTask mapTask; |
6022 /** | 6328 /** |
6023 * Any required packages that need to be fetched in order to execute this | 6329 * Any required packages that need to be fetched in order to execute |
6024 * WorkItem. | 6330 * this WorkItem. |
6025 */ | 6331 */ |
6026 core.List<Package> packages; | 6332 core.List<Package> packages; |
6027 /** Identifies the cloud project this WorkItem belongs to. */ | 6333 /** Identifies the cloud project this WorkItem belongs to. */ |
6028 core.String projectId; | 6334 core.String projectId; |
6029 /** Recommended reporting interval. */ | 6335 /** Recommended reporting interval. */ |
6030 core.String reportStatusInterval; | 6336 core.String reportStatusInterval; |
6031 /** Additional information for SeqMapTask WorkItems. */ | 6337 /** Additional information for SeqMapTask WorkItems. */ |
6032 SeqMapTask seqMapTask; | 6338 SeqMapTask seqMapTask; |
6033 /** Additional information for ShellTask WorkItems. */ | 6339 /** Additional information for ShellTask WorkItems. */ |
6034 ShellTask shellTask; | 6340 ShellTask shellTask; |
(...skipping 101 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6136 _json["streamingConfigTask"] = (streamingConfigTask).toJson(); | 6442 _json["streamingConfigTask"] = (streamingConfigTask).toJson(); |
6137 } | 6443 } |
6138 if (streamingSetupTask != null) { | 6444 if (streamingSetupTask != null) { |
6139 _json["streamingSetupTask"] = (streamingSetupTask).toJson(); | 6445 _json["streamingSetupTask"] = (streamingSetupTask).toJson(); |
6140 } | 6446 } |
6141 return _json; | 6447 return _json; |
6142 } | 6448 } |
6143 } | 6449 } |
6144 | 6450 |
6145 /** | 6451 /** |
6146 * The Dataflow service's idea of the current state of a WorkItem being | 6452 * The Dataflow service's idea of the current state of a WorkItem |
6147 * processed by a worker. | 6453 * being processed by a worker. |
6148 */ | 6454 */ |
6149 class WorkItemServiceState { | 6455 class WorkItemServiceState { |
6150 /** | 6456 /** |
6151 * Other data returned by the service, specific to the particular worker | 6457 * Other data returned by the service, specific to the particular |
6152 * harness. | 6458 * worker harness. |
6153 * | 6459 * |
6154 * The values for Object must be JSON objects. It can consist of `num`, | 6460 * The values for Object must be JSON objects. It can consist of `num`, |
6155 * `String`, `bool` and `null` as well as `Map` and `List` values. | 6461 * `String`, `bool` and `null` as well as `Map` and `List` values. |
6156 */ | 6462 */ |
6157 core.Map<core.String, core.Object> harnessData; | 6463 core.Map<core.String, core.Object> harnessData; |
6158 /** Time at which the current lease will expire. */ | 6464 /** Time at which the current lease will expire. */ |
6159 core.String leaseExpireTime; | 6465 core.String leaseExpireTime; |
6160 /** | 6466 /** |
6161 * The short ids that workers should use in subsequent metric updates. Workers | 6467 * The short ids that workers should use in subsequent metric updates. |
6162 * should strive to use short ids whenever possible, but it is ok to request | 6468 * Workers should strive to use short ids whenever possible, but it is ok |
6163 * the short_id again if a worker lost track of it (e.g. if the worker is | 6469 * to request the short_id again if a worker lost track of it |
6164 * recovering from a crash). NOTE: it is possible that the response may have | 6470 * (e.g. if the worker is recovering from a crash). |
6165 * short ids for a subset of the metrics. | 6471 * NOTE: it is possible that the response may have short ids for a subset |
| 6472 * of the metrics. |
6166 */ | 6473 */ |
6167 core.List<MetricShortId> metricShortId; | 6474 core.List<MetricShortId> metricShortId; |
6168 /** | 6475 /** |
6169 * The index value to use for the next report sent by the worker. Note: If the | 6476 * The index value to use for the next report sent by the worker. |
6170 * report call fails for whatever reason, the worker should reuse this index | 6477 * Note: If the report call fails for whatever reason, the worker should |
6171 * for subsequent report attempts. | 6478 * reuse this index for subsequent report attempts. |
6172 */ | 6479 */ |
6173 core.String nextReportIndex; | 6480 core.String nextReportIndex; |
6174 /** New recommended reporting interval. */ | 6481 /** New recommended reporting interval. */ |
6175 core.String reportStatusInterval; | 6482 core.String reportStatusInterval; |
6176 /** | 6483 /** |
6177 * The progress point in the WorkItem where the Dataflow service suggests that | 6484 * The progress point in the WorkItem where the Dataflow service |
6178 * the worker truncate the task. | 6485 * suggests that the worker truncate the task. |
6179 */ | 6486 */ |
6180 ApproximateSplitRequest splitRequest; | 6487 ApproximateSplitRequest splitRequest; |
6181 /** DEPRECATED in favor of split_request. */ | 6488 /** DEPRECATED in favor of split_request. */ |
6182 ApproximateProgress suggestedStopPoint; | 6489 ApproximateProgress suggestedStopPoint; |
6183 /** Obsolete, always empty. */ | 6490 /** Obsolete, always empty. */ |
6184 Position suggestedStopPosition; | 6491 Position suggestedStopPosition; |
6185 | 6492 |
6186 WorkItemServiceState(); | 6493 WorkItemServiceState(); |
6187 | 6494 |
6188 WorkItemServiceState.fromJson(core.Map _json) { | 6495 WorkItemServiceState.fromJson(core.Map _json) { |
(...skipping 55 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6244 | 6551 |
6245 /** Conveys a worker's progress through the work described by a WorkItem. */ | 6552 /** Conveys a worker's progress through the work described by a WorkItem. */ |
6246 class WorkItemStatus { | 6553 class WorkItemStatus { |
6247 /** True if the WorkItem was completed (successfully or unsuccessfully). */ | 6554 /** True if the WorkItem was completed (successfully or unsuccessfully). */ |
6248 core.bool completed; | 6555 core.bool completed; |
6249 /** Worker output counters for this WorkItem. */ | 6556 /** Worker output counters for this WorkItem. */ |
6250 core.List<CounterUpdate> counterUpdates; | 6557 core.List<CounterUpdate> counterUpdates; |
6251 /** See documentation of stop_position. */ | 6558 /** See documentation of stop_position. */ |
6252 DynamicSourceSplit dynamicSourceSplit; | 6559 DynamicSourceSplit dynamicSourceSplit; |
6253 /** | 6560 /** |
6254 * Specifies errors which occurred during processing. If errors are provided, | 6561 * Specifies errors which occurred during processing. If errors are |
6255 * and completed = true, then the WorkItem is considered to have failed. | 6562 * provided, and completed = true, then the WorkItem is considered |
| 6563 * to have failed. |
6256 */ | 6564 */ |
6257 core.List<Status> errors; | 6565 core.List<Status> errors; |
6258 /** DEPRECATED in favor of counter_updates. */ | 6566 /** DEPRECATED in favor of counter_updates. */ |
6259 core.List<MetricUpdate> metricUpdates; | 6567 core.List<MetricUpdate> metricUpdates; |
6260 /** DEPRECATED in favor of reported_progress. */ | 6568 /** DEPRECATED in favor of reported_progress. */ |
6261 ApproximateProgress progress; | 6569 ApproximateProgress progress; |
6262 /** | 6570 /** |
6263 * The report index. When a WorkItem is leased, the lease will contain an | 6571 * The report index. When a WorkItem is leased, the lease will |
6264 * initial report index. When a WorkItem's status is reported to the system, | 6572 * contain an initial report index. When a WorkItem's status is |
6265 * the report should be sent with that report index, and the response will | 6573 * reported to the system, the report should be sent with |
6266 * contain the index the worker should use for the next report. Reports | 6574 * that report index, and the response will contain the index the |
6267 * received with unexpected index values will be rejected by the service. In | 6575 * worker should use for the next report. Reports received with |
6268 * order to preserve idempotency, the worker should not alter the contents of | 6576 * unexpected index values will be rejected by the service. |
6269 * a report, even if the worker must submit the same report multiple times | 6577 * |
6270 * before getting back a response. The worker should not submit a subsequent | 6578 * In order to preserve idempotency, the worker should not alter the |
6271 * report until the response for the previous report had been received from | 6579 * contents of a report, even if the worker must submit the same |
6272 * the service. | 6580 * report multiple times before getting back a response. The worker |
| 6581 * should not submit a subsequent report until the response for the |
| 6582 * previous report had been received from the service. |
6273 */ | 6583 */ |
6274 core.String reportIndex; | 6584 core.String reportIndex; |
6275 /** The worker's progress through this WorkItem. */ | 6585 /** The worker's progress through this WorkItem. */ |
6276 ApproximateReportedProgress reportedProgress; | 6586 ApproximateReportedProgress reportedProgress; |
6277 /** Amount of time the worker requests for its lease. */ | 6587 /** Amount of time the worker requests for its lease. */ |
6278 core.String requestedLeaseDuration; | 6588 core.String requestedLeaseDuration; |
6279 /** DEPRECATED in favor of dynamic_source_split. */ | 6589 /** DEPRECATED in favor of dynamic_source_split. */ |
6280 SourceFork sourceFork; | 6590 SourceFork sourceFork; |
6281 /** | 6591 /** |
6282 * If the work item represented a SourceOperationRequest, and the work is | 6592 * If the work item represented a SourceOperationRequest, and the work |
6283 * completed, contains the result of the operation. | 6593 * is completed, contains the result of the operation. |
6284 */ | 6594 */ |
6285 SourceOperationResponse sourceOperationResponse; | 6595 SourceOperationResponse sourceOperationResponse; |
6286 /** | 6596 /** |
6287 * A worker may split an active map task in two parts, "primary" and | 6597 * A worker may split an active map task in two parts, "primary" and |
6288 * "residual", continuing to process the primary part and returning the | 6598 * "residual", continuing to process the primary part and returning the |
6289 * residual part into the pool of available work. This event is called a | 6599 * residual part into the pool of available work. |
6290 * "dynamic split" and is critical to the dynamic work rebalancing feature. | 6600 * This event is called a "dynamic split" and is critical to the dynamic |
6291 * The two obtained sub-tasks are called "parts" of the split. The parts, if | 6601 * work rebalancing feature. The two obtained sub-tasks are called |
6292 * concatenated, must represent the same input as would be read by the current | 6602 * "parts" of the split. |
6293 * task if the split did not happen. The exact way in which the original task | 6603 * The parts, if concatenated, must represent the same input as would |
6294 * is decomposed into the two parts is specified either as a position | 6604 * be read by the current task if the split did not happen. |
6295 * demarcating them (stop_position), or explicitly as two DerivedSources, if | 6605 * The exact way in which the original task is decomposed into the two |
6296 * this task consumes a user-defined source type (dynamic_source_split). The | 6606 * parts is specified either as a position demarcating them |
6297 * "current" task is adjusted as a result of the split: after a task with | 6607 * (stop_position), or explicitly as two DerivedSources, if this |
6298 * range [A, B) sends a stop_position update at C, its range is considered to | 6608 * task consumes a user-defined source type (dynamic_source_split). |
6299 * be [A, C), e.g.: * Progress should be interpreted relative to the new | 6609 * |
6300 * range, e.g. "75% completed" means "75% of [A, C) completed" * The worker | 6610 * The "current" task is adjusted as a result of the split: after a task |
6301 * should interpret proposed_stop_position relative to the new range, e.g. | 6611 * with range [A, B) sends a stop_position update at C, its range is |
6302 * "split at 68%" should be interpreted as "split at 68% of [A, C)". * If the | 6612 * considered to be [A, C), e.g.: |
6303 * worker chooses to split again using stop_position, only stop_positions in | 6613 * * Progress should be interpreted relative to the new range, e.g. |
6304 * [A, C) will be accepted. * Etc. dynamic_source_split has similar semantics: | 6614 * "75% completed" means "75% of [A, C) completed" |
6305 * e.g., if a task with source S splits using dynamic_source_split into {P, R} | 6615 * * The worker should interpret proposed_stop_position relative to the |
6306 * (where P and R must be together equivalent to S), then subsequent progress | 6616 * new range, e.g. "split at 68%" should be interpreted as |
6307 * and proposed_stop_position should be interpreted relative to P, and in a | 6617 * "split at 68% of [A, C)". |
6308 * potential subsequent dynamic_source_split into {P', R'}, P' and R' must be | 6618 * * If the worker chooses to split again using stop_position, only |
6309 * together equivalent to P, etc. | 6619 * stop_positions in [A, C) will be accepted. |
| 6620 * * Etc. |
| 6621 * dynamic_source_split has similar semantics: e.g., if a task with |
| 6622 * source S splits using dynamic_source_split into {P, R} |
| 6623 * (where P and R must be together equivalent to S), then subsequent |
| 6624 * progress and proposed_stop_position should be interpreted relative |
| 6625 * to P, and in a potential subsequent dynamic_source_split into {P', R'}, |
| 6626 * P' and R' must be together equivalent to P, etc. |
6310 */ | 6627 */ |
6311 Position stopPosition; | 6628 Position stopPosition; |
6312 /** Identifies the WorkItem. */ | 6629 /** Identifies the WorkItem. */ |
6313 core.String workItemId; | 6630 core.String workItemId; |
6314 | 6631 |
6315 WorkItemStatus(); | 6632 WorkItemStatus(); |
6316 | 6633 |
6317 WorkItemStatus.fromJson(core.Map _json) { | 6634 WorkItemStatus.fromJson(core.Map _json) { |
6318 if (_json.containsKey("completed")) { | 6635 if (_json.containsKey("completed")) { |
6319 completed = _json["completed"]; | 6636 completed = _json["completed"]; |
(...skipping 75 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6395 _json["stopPosition"] = (stopPosition).toJson(); | 6712 _json["stopPosition"] = (stopPosition).toJson(); |
6396 } | 6713 } |
6397 if (workItemId != null) { | 6714 if (workItemId != null) { |
6398 _json["workItemId"] = workItemId; | 6715 _json["workItemId"] = workItemId; |
6399 } | 6716 } |
6400 return _json; | 6717 return _json; |
6401 } | 6718 } |
6402 } | 6719 } |
6403 | 6720 |
6404 /** | 6721 /** |
6405 * WorkerHealthReport contains information about the health of a worker. The VM | 6722 * WorkerHealthReport contains information about the health of a worker. |
6406 * should be identified by the labels attached to the WorkerMessage that this | 6723 * |
6407 * health ping belongs to. | 6724 * The VM should be identified by the labels attached to the WorkerMessage that |
| 6725 * this health ping belongs to. |
6408 */ | 6726 */ |
6409 class WorkerHealthReport { | 6727 class WorkerHealthReport { |
6410 /** | 6728 /** |
6411 * The pods running on the worker. See: | 6729 * The pods running on the worker. See: |
6412 * http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html#_v1_pod | 6730 * http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html#_v1_pod |
| 6731 * |
6413 * This field is used by the worker to send the status of the indvidual | 6732 * This field is used by the worker to send the status of the indvidual |
6414 * containers running on each worker. | 6733 * containers running on each worker. |
6415 * | 6734 * |
6416 * The values for Object must be JSON objects. It can consist of `num`, | 6735 * The values for Object must be JSON objects. It can consist of `num`, |
6417 * `String`, `bool` and `null` as well as `Map` and `List` values. | 6736 * `String`, `bool` and `null` as well as `Map` and `List` values. |
6418 */ | 6737 */ |
6419 core.List<core.Map<core.String, core.Object>> pods; | 6738 core.List<core.Map<core.String, core.Object>> pods; |
6420 /** | 6739 /** |
6421 * The interval at which the worker is sending health reports. The default | 6740 * The interval at which the worker is sending health reports. |
6422 * value of 0 should be interpreted as the field is not being explicitly set | 6741 * The default value of 0 should be interpreted as the field is not being |
6423 * by the worker. | 6742 * explicitly set by the worker. |
6424 */ | 6743 */ |
6425 core.String reportInterval; | 6744 core.String reportInterval; |
6426 /** Whether the VM is healthy. */ | 6745 /** Whether the VM is healthy. */ |
6427 core.bool vmIsHealthy; | 6746 core.bool vmIsHealthy; |
6428 /** The time the VM was booted. */ | 6747 /** The time the VM was booted. */ |
6429 core.String vmStartupTime; | 6748 core.String vmStartupTime; |
6430 | 6749 |
6431 WorkerHealthReport(); | 6750 WorkerHealthReport(); |
6432 | 6751 |
6433 WorkerHealthReport.fromJson(core.Map _json) { | 6752 WorkerHealthReport.fromJson(core.Map _json) { |
(...skipping 23 matching lines...) Expand all Loading... |
6457 _json["vmIsHealthy"] = vmIsHealthy; | 6776 _json["vmIsHealthy"] = vmIsHealthy; |
6458 } | 6777 } |
6459 if (vmStartupTime != null) { | 6778 if (vmStartupTime != null) { |
6460 _json["vmStartupTime"] = vmStartupTime; | 6779 _json["vmStartupTime"] = vmStartupTime; |
6461 } | 6780 } |
6462 return _json; | 6781 return _json; |
6463 } | 6782 } |
6464 } | 6783 } |
6465 | 6784 |
6466 /** | 6785 /** |
6467 * WorkerHealthReportResponse contains information returned to the worker in | 6786 * WorkerHealthReportResponse contains information returned to the worker |
6468 * response to a health ping. | 6787 * in response to a health ping. |
6469 */ | 6788 */ |
6470 class WorkerHealthReportResponse { | 6789 class WorkerHealthReportResponse { |
6471 /** | 6790 /** |
6472 * A positive value indicates the worker should change its reporting interval | 6791 * A positive value indicates the worker should change its reporting interval |
6473 * to the specified value. The default value of zero means no change in report | 6792 * to the specified value. |
6474 * rate is requested by the server. | 6793 * |
| 6794 * The default value of zero means no change in report rate is requested by |
| 6795 * the server. |
6475 */ | 6796 */ |
6476 core.String reportInterval; | 6797 core.String reportInterval; |
6477 | 6798 |
6478 WorkerHealthReportResponse(); | 6799 WorkerHealthReportResponse(); |
6479 | 6800 |
6480 WorkerHealthReportResponse.fromJson(core.Map _json) { | 6801 WorkerHealthReportResponse.fromJson(core.Map _json) { |
6481 if (_json.containsKey("reportInterval")) { | 6802 if (_json.containsKey("reportInterval")) { |
6482 reportInterval = _json["reportInterval"]; | 6803 reportInterval = _json["reportInterval"]; |
6483 } | 6804 } |
6484 } | 6805 } |
6485 | 6806 |
6486 core.Map toJson() { | 6807 core.Map toJson() { |
6487 var _json = new core.Map(); | 6808 var _json = new core.Map(); |
6488 if (reportInterval != null) { | 6809 if (reportInterval != null) { |
6489 _json["reportInterval"] = reportInterval; | 6810 _json["reportInterval"] = reportInterval; |
6490 } | 6811 } |
6491 return _json; | 6812 return _json; |
6492 } | 6813 } |
6493 } | 6814 } |
6494 | 6815 |
6495 /** WorkerMessage provides information to the backend about a worker. */ | 6816 /** WorkerMessage provides information to the backend about a worker. */ |
6496 class WorkerMessage { | 6817 class WorkerMessage { |
6497 /** | 6818 /** |
6498 * Labels are used to group WorkerMessages. For example, a worker_message | 6819 * Labels are used to group WorkerMessages. |
6499 * about a particular container might have the labels: { "JOB_ID": | 6820 * For example, a worker_message about a particular container |
6500 * "2015-04-22", "WORKER_ID": "wordcount-vm-2015…" "CONTAINER_TYPE": "worker", | 6821 * might have the labels: |
6501 * "CONTAINER_ID": "ac1234def"} Label tags typically correspond to Label enum | 6822 * { "JOB_ID": "2015-04-22", |
6502 * values. However, for ease of development other strings can be used as tags. | 6823 * "WORKER_ID": "wordcount-vm-2015…" |
6503 * LABEL_UNSPECIFIED should not be used here. | 6824 * "CONTAINER_TYPE": "worker", |
| 6825 * "CONTAINER_ID": "ac1234def"} |
| 6826 * Label tags typically correspond to Label enum values. However, for ease |
| 6827 * of development other strings can be used as tags. LABEL_UNSPECIFIED should |
| 6828 * not be used here. |
6504 */ | 6829 */ |
6505 core.Map<core.String, core.String> labels; | 6830 core.Map<core.String, core.String> labels; |
6506 /** The timestamp of the worker_message. */ | 6831 /** The timestamp of the worker_message. */ |
6507 core.String time; | 6832 core.String time; |
6508 /** The health of a worker. */ | 6833 /** The health of a worker. */ |
6509 WorkerHealthReport workerHealthReport; | 6834 WorkerHealthReport workerHealthReport; |
6510 /** A worker message code. */ | 6835 /** A worker message code. */ |
6511 WorkerMessageCode workerMessageCode; | 6836 WorkerMessageCode workerMessageCode; |
6512 /** Resource metrics reported by workers. */ | 6837 /** Resource metrics reported by workers. */ |
6513 ResourceUtilizationReport workerMetrics; | 6838 ResourceUtilizationReport workerMetrics; |
(...skipping 34 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6548 } | 6873 } |
6549 if (workerMetrics != null) { | 6874 if (workerMetrics != null) { |
6550 _json["workerMetrics"] = (workerMetrics).toJson(); | 6875 _json["workerMetrics"] = (workerMetrics).toJson(); |
6551 } | 6876 } |
6552 return _json; | 6877 return _json; |
6553 } | 6878 } |
6554 } | 6879 } |
6555 | 6880 |
6556 /** | 6881 /** |
6557 * A message code is used to report status and error messages to the service. | 6882 * A message code is used to report status and error messages to the service. |
6558 * The message codes are intended to be machine readable. The service will take | 6883 * The message codes are intended to be machine readable. The service will |
6559 * care of translating these into user understandable messages if necessary. | 6884 * take care of translating these into user understandable messages if |
6560 * Example use cases: 1. Worker processes reporting successful startup. 2. | 6885 * necessary. |
6561 * Worker processes reporting specific errors (e.g. package staging failure). | 6886 * |
| 6887 * Example use cases: |
| 6888 * 1. Worker processes reporting successful startup. |
| 6889 * 2. Worker processes reporting specific errors (e.g. package staging |
| 6890 * failure). |
6562 */ | 6891 */ |
6563 class WorkerMessageCode { | 6892 class WorkerMessageCode { |
6564 /** | 6893 /** |
6565 * The code is a string intended for consumption by a machine that identifies | 6894 * The code is a string intended for consumption by a machine that identifies |
6566 * the type of message being sent. Examples: 1. "HARNESS_STARTED" might be | 6895 * the type of message being sent. |
6567 * used to indicate the worker harness has started. 2. "GCS_DOWNLOAD_ERROR" | 6896 * Examples: |
6568 * might be used to indicate an error downloading a GCS file as part of the | 6897 * 1. "HARNESS_STARTED" might be used to indicate the worker harness has |
6569 * boot process of one of the worker containers. This is a string and not an | 6898 * started. |
6570 * enum to make it easy to add new codes without waiting for an API change. | 6899 * 2. "GCS_DOWNLOAD_ERROR" might be used to indicate an error downloading |
| 6900 * a GCS file as part of the boot process of one of the worker containers. |
| 6901 * |
| 6902 * This is a string and not an enum to make it easy to add new codes without |
| 6903 * waiting for an API change. |
6571 */ | 6904 */ |
6572 core.String code; | 6905 core.String code; |
6573 /** | 6906 /** |
6574 * Parameters contains specific information about the code. This is a struct | 6907 * Parameters contains specific information about the code. |
6575 * to allow parameters of different types. Examples: 1. For a | 6908 * |
6576 * "HARNESS_STARTED" message parameters might provide the name of the worker | 6909 * This is a struct to allow parameters of different types. |
6577 * and additional data like timing information. 2. For a "GCS_DOWNLOAD_ERROR" | 6910 * |
6578 * parameters might contain fields listing the GCS objects being downloaded | 6911 * Examples: |
6579 * and fields containing errors. In general complex data structures should be | 6912 * 1. For a "HARNESS_STARTED" message parameters might provide the name |
6580 * avoided. If a worker needs to send a specific and complicated data | 6913 * of the worker and additional data like timing information. |
6581 * structure then please consider defining a new proto and adding it to the | 6914 * 2. For a "GCS_DOWNLOAD_ERROR" parameters might contain fields listing |
6582 * data oneof in WorkerMessageResponse. Conventions: Parameters should only be | 6915 * the GCS objects being downloaded and fields containing errors. |
6583 * used for information that isn't typically passed as a label. hostname and | 6916 * |
6584 * other worker identifiers should almost always be passed as labels since | 6917 * In general complex data structures should be avoided. If a worker |
6585 * they will be included on most messages. | 6918 * needs to send a specific and complicated data structure then please |
| 6919 * consider defining a new proto and adding it to the data oneof in |
| 6920 * WorkerMessageResponse. |
| 6921 * |
| 6922 * Conventions: |
| 6923 * Parameters should only be used for information that isn't typically passed |
| 6924 * as a label. |
| 6925 * hostname and other worker identifiers should almost always be passed |
| 6926 * as labels since they will be included on most messages. |
6586 * | 6927 * |
6587 * The values for Object must be JSON objects. It can consist of `num`, | 6928 * The values for Object must be JSON objects. It can consist of `num`, |
6588 * `String`, `bool` and `null` as well as `Map` and `List` values. | 6929 * `String`, `bool` and `null` as well as `Map` and `List` values. |
6589 */ | 6930 */ |
6590 core.Map<core.String, core.Object> parameters; | 6931 core.Map<core.String, core.Object> parameters; |
6591 | 6932 |
6592 WorkerMessageCode(); | 6933 WorkerMessageCode(); |
6593 | 6934 |
6594 WorkerMessageCode.fromJson(core.Map _json) { | 6935 WorkerMessageCode.fromJson(core.Map _json) { |
6595 if (_json.containsKey("code")) { | 6936 if (_json.containsKey("code")) { |
(...skipping 43 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6639 _json["workerHealthReportResponse"] = (workerHealthReportResponse).toJson(
); | 6980 _json["workerHealthReportResponse"] = (workerHealthReportResponse).toJson(
); |
6640 } | 6981 } |
6641 if (workerMetricsResponse != null) { | 6982 if (workerMetricsResponse != null) { |
6642 _json["workerMetricsResponse"] = (workerMetricsResponse).toJson(); | 6983 _json["workerMetricsResponse"] = (workerMetricsResponse).toJson(); |
6643 } | 6984 } |
6644 return _json; | 6985 return _json; |
6645 } | 6986 } |
6646 } | 6987 } |
6647 | 6988 |
6648 /** | 6989 /** |
6649 * Describes one particular pool of Cloud Dataflow workers to be instantiated by | 6990 * Describes one particular pool of Cloud Dataflow workers to be |
6650 * the Cloud Dataflow service in order to perform the computations required by a | 6991 * instantiated by the Cloud Dataflow service in order to perform the |
6651 * job. Note that a workflow job may use multiple pools, in order to match the | 6992 * computations required by a job. Note that a workflow job may use |
6652 * various computational requirements of the various stages of the job. | 6993 * multiple pools, in order to match the various computational |
| 6994 * requirements of the various stages of the job. |
6653 */ | 6995 */ |
6654 class WorkerPool { | 6996 class WorkerPool { |
6655 /** Settings for autoscaling of this WorkerPool. */ | 6997 /** Settings for autoscaling of this WorkerPool. */ |
6656 AutoscalingSettings autoscalingSettings; | 6998 AutoscalingSettings autoscalingSettings; |
6657 /** Data disks that are used by a VM in this workflow. */ | 6999 /** Data disks that are used by a VM in this workflow. */ |
6658 core.List<Disk> dataDisks; | 7000 core.List<Disk> dataDisks; |
6659 /** | 7001 /** |
6660 * The default package set to install. This allows the service to select a | 7002 * The default package set to install. This allows the service to |
6661 * default set of packages which are useful to worker harnesses written in a | 7003 * select a default set of packages which are useful to worker |
6662 * particular language. | 7004 * harnesses written in a particular language. |
6663 * Possible string values are: | 7005 * Possible string values are: |
6664 * - "DEFAULT_PACKAGE_SET_UNKNOWN" : A DEFAULT_PACKAGE_SET_UNKNOWN. | 7006 * - "DEFAULT_PACKAGE_SET_UNKNOWN" : The default set of packages to stage is |
6665 * - "DEFAULT_PACKAGE_SET_NONE" : A DEFAULT_PACKAGE_SET_NONE. | 7007 * unknown, or unspecified. |
6666 * - "DEFAULT_PACKAGE_SET_JAVA" : A DEFAULT_PACKAGE_SET_JAVA. | 7008 * - "DEFAULT_PACKAGE_SET_NONE" : Indicates that no packages should be staged |
6667 * - "DEFAULT_PACKAGE_SET_PYTHON" : A DEFAULT_PACKAGE_SET_PYTHON. | 7009 * at the worker unless |
| 7010 * explicitly specified by the job. |
| 7011 * - "DEFAULT_PACKAGE_SET_JAVA" : Stage packages typically useful to workers |
| 7012 * written in Java. |
| 7013 * - "DEFAULT_PACKAGE_SET_PYTHON" : Stage pacakges typically useful to workers |
| 7014 * written in Python. |
6668 */ | 7015 */ |
6669 core.String defaultPackageSet; | 7016 core.String defaultPackageSet; |
6670 /** | 7017 /** |
6671 * Size of root disk for VMs, in GB. If zero or unspecified, the service will | 7018 * Size of root disk for VMs, in GB. If zero or unspecified, the service will |
6672 * attempt to choose a reasonable default. | 7019 * attempt to choose a reasonable default. |
6673 */ | 7020 */ |
6674 core.int diskSizeGb; | 7021 core.int diskSizeGb; |
6675 /** Fully qualified source image for disks. */ | 7022 /** Fully qualified source image for disks. */ |
6676 core.String diskSourceImage; | 7023 core.String diskSourceImage; |
6677 /** | 7024 /** |
6678 * Type of root disk for VMs. If empty or unspecified, the service will | 7025 * Type of root disk for VMs. If empty or unspecified, the service will |
6679 * attempt to choose a reasonable default. | 7026 * attempt to choose a reasonable default. |
6680 */ | 7027 */ |
6681 core.String diskType; | 7028 core.String diskType; |
6682 /** | 7029 /** |
6683 * Configuration for VM IPs. | 7030 * Configuration for VM IPs. |
6684 * Possible string values are: | 7031 * Possible string values are: |
6685 * - "WORKER_IP_UNSPECIFIED" : A WORKER_IP_UNSPECIFIED. | 7032 * - "WORKER_IP_UNSPECIFIED" : The configuration is unknown, or unspecified. |
6686 * - "WORKER_IP_PUBLIC" : A WORKER_IP_PUBLIC. | 7033 * - "WORKER_IP_PUBLIC" : Workers should have public IP addresses. |
6687 * - "WORKER_IP_PRIVATE" : A WORKER_IP_PRIVATE. | 7034 * - "WORKER_IP_PRIVATE" : Workers should have private IP addresses. |
6688 */ | 7035 */ |
6689 core.String ipConfiguration; | 7036 core.String ipConfiguration; |
6690 /** | 7037 /** |
6691 * The kind of the worker pool; currently only `harness` and `shuffle` are | 7038 * The kind of the worker pool; currently only `harness` and `shuffle` |
6692 * supported. | 7039 * are supported. |
6693 */ | 7040 */ |
6694 core.String kind; | 7041 core.String kind; |
6695 /** | 7042 /** |
6696 * Machine type (e.g. "n1-standard-1"). If empty or unspecified, the service | 7043 * Machine type (e.g. "n1-standard-1"). If empty or unspecified, the |
6697 * will attempt to choose a reasonable default. | 7044 * service will attempt to choose a reasonable default. |
6698 */ | 7045 */ |
6699 core.String machineType; | 7046 core.String machineType; |
6700 /** Metadata to set on the Google Compute Engine VMs. */ | 7047 /** Metadata to set on the Google Compute Engine VMs. */ |
6701 core.Map<core.String, core.String> metadata; | 7048 core.Map<core.String, core.String> metadata; |
6702 /** | 7049 /** |
6703 * Network to which VMs will be assigned. If empty or unspecified, the service | 7050 * Network to which VMs will be assigned. If empty or unspecified, |
6704 * will use the network "default". | 7051 * the service will use the network "default". |
6705 */ | 7052 */ |
6706 core.String network; | 7053 core.String network; |
6707 /** | 7054 /** |
6708 * The number of threads per worker harness. If empty or unspecified, the | 7055 * The number of threads per worker harness. If empty or unspecified, the |
6709 * service will choose a number of threads (according to the number of cores | 7056 * service will choose a number of threads (according to the number of cores |
6710 * on the selected machine type for batch, or 1 by convention for streaming). | 7057 * on the selected machine type for batch, or 1 by convention for streaming). |
6711 */ | 7058 */ |
6712 core.int numThreadsPerWorker; | 7059 core.int numThreadsPerWorker; |
6713 /** | 7060 /** |
6714 * Number of Google Compute Engine workers in this pool needed to execute the | 7061 * Number of Google Compute Engine workers in this pool needed to |
6715 * job. If zero or unspecified, the service will attempt to choose a | 7062 * execute the job. If zero or unspecified, the service will |
6716 * reasonable default. | 7063 * attempt to choose a reasonable default. |
6717 */ | 7064 */ |
6718 core.int numWorkers; | 7065 core.int numWorkers; |
6719 /** | 7066 /** |
6720 * The action to take on host maintenance, as defined by the Google Compute | 7067 * The action to take on host maintenance, as defined by the Google |
6721 * Engine API. | 7068 * Compute Engine API. |
6722 */ | 7069 */ |
6723 core.String onHostMaintenance; | 7070 core.String onHostMaintenance; |
6724 /** Packages to be installed on workers. */ | 7071 /** Packages to be installed on workers. */ |
6725 core.List<Package> packages; | 7072 core.List<Package> packages; |
6726 /** | 7073 /** |
6727 * Extra arguments for this worker pool. | 7074 * Extra arguments for this worker pool. |
6728 * | 7075 * |
6729 * The values for Object must be JSON objects. It can consist of `num`, | 7076 * The values for Object must be JSON objects. It can consist of `num`, |
6730 * `String`, `bool` and `null` as well as `Map` and `List` values. | 7077 * `String`, `bool` and `null` as well as `Map` and `List` values. |
6731 */ | 7078 */ |
6732 core.Map<core.String, core.Object> poolArgs; | 7079 core.Map<core.String, core.Object> poolArgs; |
6733 /** | 7080 /** |
6734 * Subnetwork to which VMs will be assigned, if desired. Expected to be of the | 7081 * Subnetwork to which VMs will be assigned, if desired. Expected to be of |
6735 * form "regions/REGION/subnetworks/SUBNETWORK". | 7082 * the form "regions/REGION/subnetworks/SUBNETWORK". |
6736 */ | 7083 */ |
6737 core.String subnetwork; | 7084 core.String subnetwork; |
6738 /** | 7085 /** |
6739 * Settings passed through to Google Compute Engine workers when using the | 7086 * Settings passed through to Google Compute Engine workers when |
6740 * standard Dataflow task runner. Users should ignore this field. | 7087 * using the standard Dataflow task runner. Users should ignore |
| 7088 * this field. |
6741 */ | 7089 */ |
6742 TaskRunnerSettings taskrunnerSettings; | 7090 TaskRunnerSettings taskrunnerSettings; |
6743 /** | 7091 /** |
6744 * Sets the policy for determining when to turndown worker pool. Allowed | 7092 * Sets the policy for determining when to turndown worker pool. |
6745 * values are: `TEARDOWN_ALWAYS`, `TEARDOWN_ON_SUCCESS`, and `TEARDOWN_NEVER`. | 7093 * Allowed values are: `TEARDOWN_ALWAYS`, `TEARDOWN_ON_SUCCESS`, and |
| 7094 * `TEARDOWN_NEVER`. |
6746 * `TEARDOWN_ALWAYS` means workers are always torn down regardless of whether | 7095 * `TEARDOWN_ALWAYS` means workers are always torn down regardless of whether |
6747 * the job succeeds. `TEARDOWN_ON_SUCCESS` means workers are torn down if the | 7096 * the job succeeds. `TEARDOWN_ON_SUCCESS` means workers are torn down |
6748 * job succeeds. `TEARDOWN_NEVER` means the workers are never torn down. If | 7097 * if the job succeeds. `TEARDOWN_NEVER` means the workers are never torn |
6749 * the workers are not torn down by the service, they will continue to run and | 7098 * down. |
6750 * use Google Compute Engine VM resources in the user's project until they are | 7099 * |
6751 * explicitly terminated by the user. Because of this, Google recommends using | 7100 * If the workers are not torn down by the service, they will |
6752 * the `TEARDOWN_ALWAYS` policy except for small, manually supervised test | 7101 * continue to run and use Google Compute Engine VM resources in the |
6753 * jobs. If unknown or unspecified, the service will attempt to choose a | 7102 * user's project until they are explicitly terminated by the user. |
6754 * reasonable default. | 7103 * Because of this, Google recommends using the `TEARDOWN_ALWAYS` |
| 7104 * policy except for small, manually supervised test jobs. |
| 7105 * |
| 7106 * If unknown or unspecified, the service will attempt to choose a reasonable |
| 7107 * default. |
6755 * Possible string values are: | 7108 * Possible string values are: |
6756 * - "TEARDOWN_POLICY_UNKNOWN" : A TEARDOWN_POLICY_UNKNOWN. | 7109 * - "TEARDOWN_POLICY_UNKNOWN" : The teardown policy isn't specified, or is |
6757 * - "TEARDOWN_ALWAYS" : A TEARDOWN_ALWAYS. | 7110 * unknown. |
6758 * - "TEARDOWN_ON_SUCCESS" : A TEARDOWN_ON_SUCCESS. | 7111 * - "TEARDOWN_ALWAYS" : Always teardown the resource. |
6759 * - "TEARDOWN_NEVER" : A TEARDOWN_NEVER. | 7112 * - "TEARDOWN_ON_SUCCESS" : Teardown the resource on success. This is useful |
| 7113 * for debugging |
| 7114 * failures. |
| 7115 * - "TEARDOWN_NEVER" : Never teardown the resource. This is useful for |
| 7116 * debugging and |
| 7117 * development. |
6760 */ | 7118 */ |
6761 core.String teardownPolicy; | 7119 core.String teardownPolicy; |
6762 /** | 7120 /** |
6763 * Required. Docker container image that executes the Cloud Dataflow worker | 7121 * Required. Docker container image that executes the Cloud Dataflow worker |
6764 * harness, residing in Google Container Registry. | 7122 * harness, residing in Google Container Registry. |
6765 */ | 7123 */ |
6766 core.String workerHarnessContainerImage; | 7124 core.String workerHarnessContainerImage; |
6767 /** | 7125 /** |
6768 * Zone to run the worker pools in. If empty or unspecified, the service will | 7126 * Zone to run the worker pools in. If empty or unspecified, the service |
6769 * attempt to choose a reasonable default. | 7127 * will attempt to choose a reasonable default. |
6770 */ | 7128 */ |
6771 core.String zone; | 7129 core.String zone; |
6772 | 7130 |
6773 WorkerPool(); | 7131 WorkerPool(); |
6774 | 7132 |
6775 WorkerPool.fromJson(core.Map _json) { | 7133 WorkerPool.fromJson(core.Map _json) { |
6776 if (_json.containsKey("autoscalingSettings")) { | 7134 if (_json.containsKey("autoscalingSettings")) { |
6777 autoscalingSettings = new AutoscalingSettings.fromJson(_json["autoscalingS
ettings"]); | 7135 autoscalingSettings = new AutoscalingSettings.fromJson(_json["autoscalingS
ettings"]); |
6778 } | 7136 } |
6779 if (_json.containsKey("dataDisks")) { | 7137 if (_json.containsKey("dataDisks")) { |
(...skipping 123 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6903 if (zone != null) { | 7261 if (zone != null) { |
6904 _json["zone"] = zone; | 7262 _json["zone"] = zone; |
6905 } | 7263 } |
6906 return _json; | 7264 return _json; |
6907 } | 7265 } |
6908 } | 7266 } |
6909 | 7267 |
6910 /** Provides data to pass through to the worker harness. */ | 7268 /** Provides data to pass through to the worker harness. */ |
6911 class WorkerSettings { | 7269 class WorkerSettings { |
6912 /** | 7270 /** |
6913 * The base URL for accessing Google Cloud APIs. When workers access Google | 7271 * The base URL for accessing Google Cloud APIs. |
6914 * Cloud APIs, they logically do so via relative URLs. If this field is | 7272 * |
6915 * specified, it supplies the base URL to use for resolving these relative | 7273 * When workers access Google Cloud APIs, they logically do so via |
6916 * URLs. The normative algorithm used is defined by RFC 1808, "Relative | 7274 * relative URLs. If this field is specified, it supplies the base |
6917 * Uniform Resource Locators". If not specified, the default value is | 7275 * URL to use for resolving these relative URLs. The normative |
6918 * "http://www.googleapis.com/" | 7276 * algorithm used is defined by RFC 1808, "Relative Uniform Resource |
| 7277 * Locators". |
| 7278 * |
| 7279 * If not specified, the default value is "http://www.googleapis.com/" |
6919 */ | 7280 */ |
6920 core.String baseUrl; | 7281 core.String baseUrl; |
6921 /** Whether to send work progress updates to the service. */ | 7282 /** Whether to send work progress updates to the service. */ |
6922 core.bool reportingEnabled; | 7283 core.bool reportingEnabled; |
6923 /** | 7284 /** |
6924 * The Cloud Dataflow service path relative to the root URL, for example, | 7285 * The Cloud Dataflow service path relative to the root URL, for example, |
6925 * "dataflow/v1b3/projects". | 7286 * "dataflow/v1b3/projects". |
6926 */ | 7287 */ |
6927 core.String servicePath; | 7288 core.String servicePath; |
6928 /** | 7289 /** |
6929 * The Shuffle service path relative to the root URL, for example, | 7290 * The Shuffle service path relative to the root URL, for example, |
6930 * "shuffle/v1beta1". | 7291 * "shuffle/v1beta1". |
6931 */ | 7292 */ |
6932 core.String shuffleServicePath; | 7293 core.String shuffleServicePath; |
6933 /** | 7294 /** |
6934 * The prefix of the resources the system should use for temporary storage. | 7295 * The prefix of the resources the system should use for temporary |
6935 * The supported resource type is: Google Cloud Storage: | 7296 * storage. |
6936 * storage.googleapis.com/{bucket}/{object} | 7297 * |
6937 * bucket.storage.googleapis.com/{object} | 7298 * The supported resource type is: |
| 7299 * |
| 7300 * Google Cloud Storage: |
| 7301 * |
| 7302 * storage.googleapis.com/{bucket}/{object} |
| 7303 * bucket.storage.googleapis.com/{object} |
6938 */ | 7304 */ |
6939 core.String tempStoragePrefix; | 7305 core.String tempStoragePrefix; |
6940 /** The ID of the worker running this pipeline. */ | 7306 /** The ID of the worker running this pipeline. */ |
6941 core.String workerId; | 7307 core.String workerId; |
6942 | 7308 |
6943 WorkerSettings(); | 7309 WorkerSettings(); |
6944 | 7310 |
6945 WorkerSettings.fromJson(core.Map _json) { | 7311 WorkerSettings.fromJson(core.Map _json) { |
6946 if (_json.containsKey("baseUrl")) { | 7312 if (_json.containsKey("baseUrl")) { |
6947 baseUrl = _json["baseUrl"]; | 7313 baseUrl = _json["baseUrl"]; |
(...skipping 33 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
6981 _json["tempStoragePrefix"] = tempStoragePrefix; | 7347 _json["tempStoragePrefix"] = tempStoragePrefix; |
6982 } | 7348 } |
6983 if (workerId != null) { | 7349 if (workerId != null) { |
6984 _json["workerId"] = workerId; | 7350 _json["workerId"] = workerId; |
6985 } | 7351 } |
6986 return _json; | 7352 return _json; |
6987 } | 7353 } |
6988 } | 7354 } |
6989 | 7355 |
6990 /** | 7356 /** |
6991 * An instruction that writes records. Takes one input, produces no outputs. | 7357 * An instruction that writes records. |
| 7358 * Takes one input, produces no outputs. |
6992 */ | 7359 */ |
6993 class WriteInstruction { | 7360 class WriteInstruction { |
6994 /** The input. */ | 7361 /** The input. */ |
6995 InstructionInput input; | 7362 InstructionInput input; |
6996 /** The sink to write to. */ | 7363 /** The sink to write to. */ |
6997 Sink sink; | 7364 Sink sink; |
6998 | 7365 |
6999 WriteInstruction(); | 7366 WriteInstruction(); |
7000 | 7367 |
7001 WriteInstruction.fromJson(core.Map _json) { | 7368 WriteInstruction.fromJson(core.Map _json) { |
7002 if (_json.containsKey("input")) { | 7369 if (_json.containsKey("input")) { |
7003 input = new InstructionInput.fromJson(_json["input"]); | 7370 input = new InstructionInput.fromJson(_json["input"]); |
7004 } | 7371 } |
7005 if (_json.containsKey("sink")) { | 7372 if (_json.containsKey("sink")) { |
7006 sink = new Sink.fromJson(_json["sink"]); | 7373 sink = new Sink.fromJson(_json["sink"]); |
7007 } | 7374 } |
7008 } | 7375 } |
7009 | 7376 |
7010 core.Map toJson() { | 7377 core.Map toJson() { |
7011 var _json = new core.Map(); | 7378 var _json = new core.Map(); |
7012 if (input != null) { | 7379 if (input != null) { |
7013 _json["input"] = (input).toJson(); | 7380 _json["input"] = (input).toJson(); |
7014 } | 7381 } |
7015 if (sink != null) { | 7382 if (sink != null) { |
7016 _json["sink"] = (sink).toJson(); | 7383 _json["sink"] = (sink).toJson(); |
7017 } | 7384 } |
7018 return _json; | 7385 return _json; |
7019 } | 7386 } |
7020 } | 7387 } |
OLD | NEW |