| OLD | NEW |
| 1 // This is a generated file (see the discoveryapis_generator project). | 1 // This is a generated file (see the discoveryapis_generator project). |
| 2 | 2 |
| 3 library googleapis_beta.dataproc.v1beta1; | 3 library googleapis_beta.dataproc.v1beta1; |
| 4 | 4 |
| 5 import 'dart:core' as core; | 5 import 'dart:core' as core; |
| 6 import 'dart:async' as async; | 6 import 'dart:async' as async; |
| 7 import 'dart:convert' as convert; | 7 import 'dart:convert' as convert; |
| 8 | 8 |
| 9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons; | 9 import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons; |
| 10 import 'package:http/http.dart' as http; | 10 import 'package:http/http.dart' as http; |
| 11 | 11 |
| 12 export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show | 12 export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show |
| 13 ApiRequestError, DetailedApiRequestError; | 13 ApiRequestError, DetailedApiRequestError; |
| 14 | 14 |
| 15 const core.String USER_AGENT = 'dart-api-client dataproc/v1beta1'; | 15 const core.String USER_AGENT = 'dart-api-client dataproc/v1beta1'; |
| 16 | 16 |
| 17 /** | 17 /** Manages Hadoop-based clusters and jobs on Google Cloud Platform. */ |
| 18 * An API for managing Hadoop-based clusters and jobs on Google Cloud Platform. | |
| 19 */ | |
| 20 class DataprocApi { | 18 class DataprocApi { |
| 21 /** View and manage your data across Google Cloud Platform services */ | 19 /** View and manage your data across Google Cloud Platform services */ |
| 22 static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platf
orm"; | 20 static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platf
orm"; |
| 23 | 21 |
| 24 | 22 |
| 25 final commons.ApiRequester _requester; | 23 final commons.ApiRequester _requester; |
| 26 | 24 |
| 27 OperationsResourceApi get operations => new OperationsResourceApi(_requester); | 25 OperationsResourceApi get operations => new OperationsResourceApi(_requester); |
| 28 ProjectsResourceApi get projects => new ProjectsResourceApi(_requester); | 26 ProjectsResourceApi get projects => new ProjectsResourceApi(_requester); |
| 29 | 27 |
| 30 DataprocApi(http.Client client, {core.String rootUrl: "https://dataproc.google
apis.com/", core.String servicePath: ""}) : | 28 DataprocApi(http.Client client, {core.String rootUrl: "https://dataproc.google
apis.com/", core.String servicePath: ""}) : |
| 31 _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_A
GENT); | 29 _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_A
GENT); |
| 32 } | 30 } |
| 33 | 31 |
| 34 | 32 |
| 35 class OperationsResourceApi { | 33 class OperationsResourceApi { |
| 36 final commons.ApiRequester _requester; | 34 final commons.ApiRequester _requester; |
| 37 | 35 |
| 38 OperationsResourceApi(commons.ApiRequester client) : | 36 OperationsResourceApi(commons.ApiRequester client) : |
| 39 _requester = client; | 37 _requester = client; |
| 40 | 38 |
| 41 /** | 39 /** |
| 42 * Starts asynchronous cancellation on a long-running operation. The server | 40 * Starts asynchronous cancellation on a long-running operation. The server |
| 43 * makes a best effort to cancel the operation, but success is not guaranteed. | 41 * makes a best effort to cancel the operation, but success is not guaranteed. |
| 44 * If the server doesn't support this method, it returns | 42 * If the server doesn't support this method, it returns |
| 45 * `google.rpc.Code.UNIMPLEMENTED`. Clients can use | 43 * google.rpc.Code.UNIMPLEMENTED. Clients can use operations.get or other |
| 46 * [operations.get](/dataproc/reference/rest/v1beta1/operations/get) or other | |
| 47 * methods to check whether the cancellation succeeded or whether the | 44 * methods to check whether the cancellation succeeded or whether the |
| 48 * operation completed despite cancellation. | 45 * operation completed despite cancellation. |
| 49 * | 46 * |
| 50 * [request] - The metadata request object. | 47 * [request] - The metadata request object. |
| 51 * | 48 * |
| 52 * Request parameters: | 49 * Request parameters: |
| 53 * | 50 * |
| 54 * [name] - The name of the operation resource to be cancelled. | 51 * [name] - The name of the operation resource to be cancelled. |
| 55 * Value must have pattern "^operations/.+$". | 52 * Value must have pattern "^operations/.+$". |
| 56 * | 53 * |
| (...skipping 29 matching lines...) Expand all Loading... |
| 86 uploadOptions: _uploadOptions, | 83 uploadOptions: _uploadOptions, |
| 87 uploadMedia: _uploadMedia, | 84 uploadMedia: _uploadMedia, |
| 88 downloadOptions: _downloadOptions); | 85 downloadOptions: _downloadOptions); |
| 89 return _response.then((data) => new Empty.fromJson(data)); | 86 return _response.then((data) => new Empty.fromJson(data)); |
| 90 } | 87 } |
| 91 | 88 |
| 92 /** | 89 /** |
| 93 * Deletes a long-running operation. This method indicates that the client is | 90 * Deletes a long-running operation. This method indicates that the client is |
| 94 * no longer interested in the operation result. It does not cancel the | 91 * no longer interested in the operation result. It does not cancel the |
| 95 * operation. If the server doesn't support this method, it returns | 92 * operation. If the server doesn't support this method, it returns |
| 96 * `google.rpc.Code.UNIMPLEMENTED`. | 93 * google.rpc.Code.UNIMPLEMENTED. |
| 97 * | 94 * |
| 98 * Request parameters: | 95 * Request parameters: |
| 99 * | 96 * |
| 100 * [name] - The name of the operation resource to be deleted. | 97 * [name] - The name of the operation resource to be deleted. |
| 101 * Value must have pattern "^operations/.+$". | 98 * Value must have pattern "^operations/.+$". |
| 102 * | 99 * |
| 103 * Completes with a [Empty]. | 100 * Completes with a [Empty]. |
| 104 * | 101 * |
| 105 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 102 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 106 * error. | 103 * error. |
| (...skipping 62 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 169 body: _body, | 166 body: _body, |
| 170 queryParams: _queryParams, | 167 queryParams: _queryParams, |
| 171 uploadOptions: _uploadOptions, | 168 uploadOptions: _uploadOptions, |
| 172 uploadMedia: _uploadMedia, | 169 uploadMedia: _uploadMedia, |
| 173 downloadOptions: _downloadOptions); | 170 downloadOptions: _downloadOptions); |
| 174 return _response.then((data) => new Operation.fromJson(data)); | 171 return _response.then((data) => new Operation.fromJson(data)); |
| 175 } | 172 } |
| 176 | 173 |
| 177 /** | 174 /** |
| 178 * Lists operations that match the specified filter in the request. If the | 175 * Lists operations that match the specified filter in the request. If the |
| 179 * server doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the | 176 * server doesn't support this method, it returns UNIMPLEMENTED.NOTE: the name |
| 180 * `name` binding below allows API services to override the binding to use | 177 * binding below allows API services to override the binding to use different |
| 181 * different resource name schemes, such as `users / * /operations`. | 178 * resource name schemes, such as users / * /operations. |
| 182 * | 179 * |
| 183 * Request parameters: | 180 * Request parameters: |
| 184 * | 181 * |
| 185 * [name] - The name of the operation collection. | 182 * [name] - The name of the operation collection. |
| 186 * Value must have pattern "^operations$". | 183 * Value must have pattern "^operations$". |
| 187 * | 184 * |
| 188 * [filter] - The standard list filter. | 185 * [pageToken] - The standard list page token. |
| 189 * | 186 * |
| 190 * [pageSize] - The standard list page size. | 187 * [pageSize] - The standard list page size. |
| 191 * | 188 * |
| 192 * [pageToken] - The standard list page token. | 189 * [filter] - The standard list filter. |
| 193 * | 190 * |
| 194 * Completes with a [ListOperationsResponse]. | 191 * Completes with a [ListOperationsResponse]. |
| 195 * | 192 * |
| 196 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 193 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 197 * error. | 194 * error. |
| 198 * | 195 * |
| 199 * If the used [http.Client] completes with an error when making a REST call, | 196 * If the used [http.Client] completes with an error when making a REST call, |
| 200 * this method will complete with the same error. | 197 * this method will complete with the same error. |
| 201 */ | 198 */ |
| 202 async.Future<ListOperationsResponse> list(core.String name, {core.String filte
r, core.int pageSize, core.String pageToken}) { | 199 async.Future<ListOperationsResponse> list(core.String name, {core.String pageT
oken, core.int pageSize, core.String filter}) { |
| 203 var _url = null; | 200 var _url = null; |
| 204 var _queryParams = new core.Map(); | 201 var _queryParams = new core.Map(); |
| 205 var _uploadMedia = null; | 202 var _uploadMedia = null; |
| 206 var _uploadOptions = null; | 203 var _uploadOptions = null; |
| 207 var _downloadOptions = commons.DownloadOptions.Metadata; | 204 var _downloadOptions = commons.DownloadOptions.Metadata; |
| 208 var _body = null; | 205 var _body = null; |
| 209 | 206 |
| 210 if (name == null) { | 207 if (name == null) { |
| 211 throw new core.ArgumentError("Parameter name is required."); | 208 throw new core.ArgumentError("Parameter name is required."); |
| 212 } | 209 } |
| 210 if (pageToken != null) { |
| 211 _queryParams["pageToken"] = [pageToken]; |
| 212 } |
| 213 if (pageSize != null) { |
| 214 _queryParams["pageSize"] = ["${pageSize}"]; |
| 215 } |
| 213 if (filter != null) { | 216 if (filter != null) { |
| 214 _queryParams["filter"] = [filter]; | 217 _queryParams["filter"] = [filter]; |
| 215 } | 218 } |
| 216 if (pageSize != null) { | |
| 217 _queryParams["pageSize"] = ["${pageSize}"]; | |
| 218 } | |
| 219 if (pageToken != null) { | |
| 220 _queryParams["pageToken"] = [pageToken]; | |
| 221 } | |
| 222 | 219 |
| 223 _url = 'v1beta1/' + commons.Escaper.ecapeVariableReserved('$name'); | 220 _url = 'v1beta1/' + commons.Escaper.ecapeVariableReserved('$name'); |
| 224 | 221 |
| 225 var _response = _requester.request(_url, | 222 var _response = _requester.request(_url, |
| 226 "GET", | 223 "GET", |
| 227 body: _body, | 224 body: _body, |
| 228 queryParams: _queryParams, | 225 queryParams: _queryParams, |
| 229 uploadOptions: _uploadOptions, | 226 uploadOptions: _uploadOptions, |
| 230 uploadMedia: _uploadMedia, | 227 uploadMedia: _uploadMedia, |
| 231 downloadOptions: _downloadOptions); | 228 downloadOptions: _downloadOptions); |
| (...skipping 20 matching lines...) Expand all Loading... |
| 252 ProjectsClustersResourceApi(commons.ApiRequester client) : | 249 ProjectsClustersResourceApi(commons.ApiRequester client) : |
| 253 _requester = client; | 250 _requester = client; |
| 254 | 251 |
| 255 /** | 252 /** |
| 256 * Creates a cluster in a project. | 253 * Creates a cluster in a project. |
| 257 * | 254 * |
| 258 * [request] - The metadata request object. | 255 * [request] - The metadata request object. |
| 259 * | 256 * |
| 260 * Request parameters: | 257 * Request parameters: |
| 261 * | 258 * |
| 262 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 259 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 263 * the cluster belongs to. | 260 * cluster belongs to. |
| 264 * | 261 * |
| 265 * Completes with a [Operation]. | 262 * Completes with a [Operation]. |
| 266 * | 263 * |
| 267 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 264 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 268 * error. | 265 * error. |
| 269 * | 266 * |
| 270 * If the used [http.Client] completes with an error when making a REST call, | 267 * If the used [http.Client] completes with an error when making a REST call, |
| 271 * this method will complete with the same error. | 268 * this method will complete with the same error. |
| 272 */ | 269 */ |
| 273 async.Future<Operation> create(Cluster request, core.String projectId) { | 270 async.Future<Operation> create(Cluster request, core.String projectId) { |
| (...skipping 21 matching lines...) Expand all Loading... |
| 295 uploadMedia: _uploadMedia, | 292 uploadMedia: _uploadMedia, |
| 296 downloadOptions: _downloadOptions); | 293 downloadOptions: _downloadOptions); |
| 297 return _response.then((data) => new Operation.fromJson(data)); | 294 return _response.then((data) => new Operation.fromJson(data)); |
| 298 } | 295 } |
| 299 | 296 |
| 300 /** | 297 /** |
| 301 * Deletes a cluster in a project. | 298 * Deletes a cluster in a project. |
| 302 * | 299 * |
| 303 * Request parameters: | 300 * Request parameters: |
| 304 * | 301 * |
| 305 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 302 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 306 * the cluster belongs to. | 303 * cluster belongs to. |
| 307 * | 304 * |
| 308 * [clusterName] - [Required] The cluster name. | 305 * [clusterName] - Required The cluster name. |
| 309 * | 306 * |
| 310 * Completes with a [Operation]. | 307 * Completes with a [Operation]. |
| 311 * | 308 * |
| 312 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 309 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 313 * error. | 310 * error. |
| 314 * | 311 * |
| 315 * If the used [http.Client] completes with an error when making a REST call, | 312 * If the used [http.Client] completes with an error when making a REST call, |
| 316 * this method will complete with the same error. | 313 * this method will complete with the same error. |
| 317 */ | 314 */ |
| 318 async.Future<Operation> delete(core.String projectId, core.String clusterName)
{ | 315 async.Future<Operation> delete(core.String projectId, core.String clusterName)
{ |
| (...skipping 18 matching lines...) Expand all Loading... |
| 337 body: _body, | 334 body: _body, |
| 338 queryParams: _queryParams, | 335 queryParams: _queryParams, |
| 339 uploadOptions: _uploadOptions, | 336 uploadOptions: _uploadOptions, |
| 340 uploadMedia: _uploadMedia, | 337 uploadMedia: _uploadMedia, |
| 341 downloadOptions: _downloadOptions); | 338 downloadOptions: _downloadOptions); |
| 342 return _response.then((data) => new Operation.fromJson(data)); | 339 return _response.then((data) => new Operation.fromJson(data)); |
| 343 } | 340 } |
| 344 | 341 |
| 345 /** | 342 /** |
| 346 * Gets cluster diagnostic information. After the operation completes, the | 343 * Gets cluster diagnostic information. After the operation completes, the |
| 347 * Operation.response field contains `DiagnoseClusterOutputLocation`. | 344 * Operation.response field contains DiagnoseClusterOutputLocation. |
| 348 * | 345 * |
| 349 * [request] - The metadata request object. | 346 * [request] - The metadata request object. |
| 350 * | 347 * |
| 351 * Request parameters: | 348 * Request parameters: |
| 352 * | 349 * |
| 353 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 350 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 354 * the cluster belongs to. | 351 * cluster belongs to. |
| 355 * | 352 * |
| 356 * [clusterName] - [Required] The cluster name. | 353 * [clusterName] - Required The cluster name. |
| 357 * | 354 * |
| 358 * Completes with a [Operation]. | 355 * Completes with a [Operation]. |
| 359 * | 356 * |
| 360 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 357 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 361 * error. | 358 * error. |
| 362 * | 359 * |
| 363 * If the used [http.Client] completes with an error when making a REST call, | 360 * If the used [http.Client] completes with an error when making a REST call, |
| 364 * this method will complete with the same error. | 361 * this method will complete with the same error. |
| 365 */ | 362 */ |
| 366 async.Future<Operation> diagnose(DiagnoseClusterRequest request, core.String p
rojectId, core.String clusterName) { | 363 async.Future<Operation> diagnose(DiagnoseClusterRequest request, core.String p
rojectId, core.String clusterName) { |
| (...skipping 24 matching lines...) Expand all Loading... |
| 391 uploadMedia: _uploadMedia, | 388 uploadMedia: _uploadMedia, |
| 392 downloadOptions: _downloadOptions); | 389 downloadOptions: _downloadOptions); |
| 393 return _response.then((data) => new Operation.fromJson(data)); | 390 return _response.then((data) => new Operation.fromJson(data)); |
| 394 } | 391 } |
| 395 | 392 |
| 396 /** | 393 /** |
| 397 * Gets the resource representation for a cluster in a project. | 394 * Gets the resource representation for a cluster in a project. |
| 398 * | 395 * |
| 399 * Request parameters: | 396 * Request parameters: |
| 400 * | 397 * |
| 401 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 398 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 402 * the cluster belongs to. | 399 * cluster belongs to. |
| 403 * | 400 * |
| 404 * [clusterName] - [Required] The cluster name. | 401 * [clusterName] - Required The cluster name. |
| 405 * | 402 * |
| 406 * Completes with a [Cluster]. | 403 * Completes with a [Cluster]. |
| 407 * | 404 * |
| 408 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 405 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 409 * error. | 406 * error. |
| 410 * | 407 * |
| 411 * If the used [http.Client] completes with an error when making a REST call, | 408 * If the used [http.Client] completes with an error when making a REST call, |
| 412 * this method will complete with the same error. | 409 * this method will complete with the same error. |
| 413 */ | 410 */ |
| 414 async.Future<Cluster> get(core.String projectId, core.String clusterName) { | 411 async.Future<Cluster> get(core.String projectId, core.String clusterName) { |
| (...skipping 21 matching lines...) Expand all Loading... |
| 436 uploadMedia: _uploadMedia, | 433 uploadMedia: _uploadMedia, |
| 437 downloadOptions: _downloadOptions); | 434 downloadOptions: _downloadOptions); |
| 438 return _response.then((data) => new Cluster.fromJson(data)); | 435 return _response.then((data) => new Cluster.fromJson(data)); |
| 439 } | 436 } |
| 440 | 437 |
| 441 /** | 438 /** |
| 442 * Lists all clusters in a project. | 439 * Lists all clusters in a project. |
| 443 * | 440 * |
| 444 * Request parameters: | 441 * Request parameters: |
| 445 * | 442 * |
| 446 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 443 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 447 * the cluster belongs to. | 444 * cluster belongs to. |
| 448 * | 445 * |
| 449 * [filter] - [Optional] A filter constraining which clusters to list. Valid | 446 * [filter] - Optional A filter constraining which clusters to list. Valid |
| 450 * filters contain label terms such as: labels.key1 = val1 AND (-labels.k2 = | 447 * filters contain label terms such as: labels.key1 = val1 AND (-labels.k2 = |
| 451 * val2 OR labels.k3 = val3) | 448 * val2 OR labels.k3 = val3) |
| 452 * | 449 * |
| 450 * [pageToken] - The standard List page token. |
| 451 * |
| 453 * [pageSize] - The standard List page size. | 452 * [pageSize] - The standard List page size. |
| 454 * | 453 * |
| 455 * [pageToken] - The standard List page token. | |
| 456 * | |
| 457 * Completes with a [ListClustersResponse]. | 454 * Completes with a [ListClustersResponse]. |
| 458 * | 455 * |
| 459 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 456 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 460 * error. | 457 * error. |
| 461 * | 458 * |
| 462 * If the used [http.Client] completes with an error when making a REST call, | 459 * If the used [http.Client] completes with an error when making a REST call, |
| 463 * this method will complete with the same error. | 460 * this method will complete with the same error. |
| 464 */ | 461 */ |
| 465 async.Future<ListClustersResponse> list(core.String projectId, {core.String fi
lter, core.int pageSize, core.String pageToken}) { | 462 async.Future<ListClustersResponse> list(core.String projectId, {core.String fi
lter, core.String pageToken, core.int pageSize}) { |
| 466 var _url = null; | 463 var _url = null; |
| 467 var _queryParams = new core.Map(); | 464 var _queryParams = new core.Map(); |
| 468 var _uploadMedia = null; | 465 var _uploadMedia = null; |
| 469 var _uploadOptions = null; | 466 var _uploadOptions = null; |
| 470 var _downloadOptions = commons.DownloadOptions.Metadata; | 467 var _downloadOptions = commons.DownloadOptions.Metadata; |
| 471 var _body = null; | 468 var _body = null; |
| 472 | 469 |
| 473 if (projectId == null) { | 470 if (projectId == null) { |
| 474 throw new core.ArgumentError("Parameter projectId is required."); | 471 throw new core.ArgumentError("Parameter projectId is required."); |
| 475 } | 472 } |
| 476 if (filter != null) { | 473 if (filter != null) { |
| 477 _queryParams["filter"] = [filter]; | 474 _queryParams["filter"] = [filter]; |
| 478 } | 475 } |
| 476 if (pageToken != null) { |
| 477 _queryParams["pageToken"] = [pageToken]; |
| 478 } |
| 479 if (pageSize != null) { | 479 if (pageSize != null) { |
| 480 _queryParams["pageSize"] = ["${pageSize}"]; | 480 _queryParams["pageSize"] = ["${pageSize}"]; |
| 481 } | 481 } |
| 482 if (pageToken != null) { | |
| 483 _queryParams["pageToken"] = [pageToken]; | |
| 484 } | |
| 485 | 482 |
| 486 _url = 'v1beta1/projects/' + commons.Escaper.ecapeVariable('$projectId') + '
/clusters'; | 483 _url = 'v1beta1/projects/' + commons.Escaper.ecapeVariable('$projectId') + '
/clusters'; |
| 487 | 484 |
| 488 var _response = _requester.request(_url, | 485 var _response = _requester.request(_url, |
| 489 "GET", | 486 "GET", |
| 490 body: _body, | 487 body: _body, |
| 491 queryParams: _queryParams, | 488 queryParams: _queryParams, |
| 492 uploadOptions: _uploadOptions, | 489 uploadOptions: _uploadOptions, |
| 493 uploadMedia: _uploadMedia, | 490 uploadMedia: _uploadMedia, |
| 494 downloadOptions: _downloadOptions); | 491 downloadOptions: _downloadOptions); |
| 495 return _response.then((data) => new ListClustersResponse.fromJson(data)); | 492 return _response.then((data) => new ListClustersResponse.fromJson(data)); |
| 496 } | 493 } |
| 497 | 494 |
| 498 /** | 495 /** |
| 499 * Updates a cluster in a project. | 496 * Updates a cluster in a project. |
| 500 * | 497 * |
| 501 * [request] - The metadata request object. | 498 * [request] - The metadata request object. |
| 502 * | 499 * |
| 503 * Request parameters: | 500 * Request parameters: |
| 504 * | 501 * |
| 505 * [projectId] - [Required] The ID of the Google Cloud Platform project the | 502 * [projectId] - Required The ID of the Google Cloud Platform project the |
| 506 * cluster belongs to. | 503 * cluster belongs to. |
| 507 * | 504 * |
| 508 * [clusterName] - [Required] The cluster name. | 505 * [clusterName] - Required The cluster name. |
| 509 * | 506 * |
| 510 * [updateMask] - [Required] Specifies the path, relative to Cluster, of the | 507 * [updateMask] - Required Specifies the path, relative to |
| 511 * field to update. For example, to change the number of workers in a cluster | 508 * <code>Cluster</code>, of the field to update. For example, to change the |
| 512 * to 5, the update_mask parameter would be specified as | 509 * number of workers in a cluster to 5, the <code>update_mask</code> parameter |
| 513 * configuration.worker_configuration.num_instances, and the `PATCH` request | 510 * would be specified as |
| 514 * body would specify the new value, as follows: { "configuration":{ | 511 * <code>configuration.worker_configuration.num_instances</code>, and the |
| 515 * "workerConfiguration":{ "numInstances":"5" } } } Similarly, to change the | 512 * PATCH request body would specify the new value, as follows: |
| 516 * number of preemptible workers in a cluster to 5, the update_mask parameter | 513 * { |
| 517 * would be config.secondary_worker_config.num_instances, and the `PATCH` | 514 * "configuration":{ |
| 518 * request body would be set as follows: { "config":{ | 515 * "workerConfiguration":{ |
| 519 * "secondaryWorkerConfig":{ "numInstances":"5" } } } Note: Currently, | 516 * "numInstances":"5" |
| 520 * config.worker_config.num_instances and | 517 * } |
| 521 * config.secondary_worker_config.num_instances are the only fields that can | 518 * } |
| 522 * be updated. | 519 * } |
| 520 * Similarly, to change the number of preemptible workers in a cluster to 5, |
| 521 * the <code>update_mask</code> parameter would be |
| 522 * <code>config.secondary_worker_config.num_instances</code>, and the PATCH |
| 523 * request body would be set as follows: |
| 524 * { |
| 525 * "config":{ |
| 526 * "secondaryWorkerConfig":{ |
| 527 * "numInstances":"5" |
| 528 * } |
| 529 * } |
| 530 * } |
| 531 * <strong>Note:</strong> Currently, |
| 532 * <code>config.worker_config.num_instances</code> and |
| 533 * <code>config.secondary_worker_config.num_instances</code> are the only |
| 534 * fields that can be updated. |
| 523 * | 535 * |
| 524 * Completes with a [Operation]. | 536 * Completes with a [Operation]. |
| 525 * | 537 * |
| 526 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 538 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 527 * error. | 539 * error. |
| 528 * | 540 * |
| 529 * If the used [http.Client] completes with an error when making a REST call, | 541 * If the used [http.Client] completes with an error when making a REST call, |
| 530 * this method will complete with the same error. | 542 * this method will complete with the same error. |
| 531 */ | 543 */ |
| 532 async.Future<Operation> patch(Cluster request, core.String projectId, core.Str
ing clusterName, {core.String updateMask}) { | 544 async.Future<Operation> patch(Cluster request, core.String projectId, core.Str
ing clusterName, {core.String updateMask}) { |
| (...skipping 33 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 566 | 578 |
| 567 | 579 |
| 568 class ProjectsJobsResourceApi { | 580 class ProjectsJobsResourceApi { |
| 569 final commons.ApiRequester _requester; | 581 final commons.ApiRequester _requester; |
| 570 | 582 |
| 571 ProjectsJobsResourceApi(commons.ApiRequester client) : | 583 ProjectsJobsResourceApi(commons.ApiRequester client) : |
| 572 _requester = client; | 584 _requester = client; |
| 573 | 585 |
| 574 /** | 586 /** |
| 575 * Starts a job cancellation request. To access the job resource after | 587 * Starts a job cancellation request. To access the job resource after |
| 576 * cancellation, call | 588 * cancellation, call jobs.list or jobs.get. |
| 577 * [jobs.list](/dataproc/reference/rest/v1beta1/projects.jobs/list) or | |
| 578 * [jobs.get](/dataproc/reference/rest/v1beta1/projects.jobs/get). | |
| 579 * | 589 * |
| 580 * [request] - The metadata request object. | 590 * [request] - The metadata request object. |
| 581 * | 591 * |
| 582 * Request parameters: | 592 * Request parameters: |
| 583 * | 593 * |
| 584 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 594 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 585 * the job belongs to. | 595 * job belongs to. |
| 586 * | 596 * |
| 587 * [jobId] - [Required] The job ID. | 597 * [jobId] - Required The job ID. |
| 588 * | 598 * |
| 589 * Completes with a [Job]. | 599 * Completes with a [Job]. |
| 590 * | 600 * |
| 591 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 601 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 592 * error. | 602 * error. |
| 593 * | 603 * |
| 594 * If the used [http.Client] completes with an error when making a REST call, | 604 * If the used [http.Client] completes with an error when making a REST call, |
| 595 * this method will complete with the same error. | 605 * this method will complete with the same error. |
| 596 */ | 606 */ |
| 597 async.Future<Job> cancel(CancelJobRequest request, core.String projectId, core
.String jobId) { | 607 async.Future<Job> cancel(CancelJobRequest request, core.String projectId, core
.String jobId) { |
| (...skipping 21 matching lines...) Expand all Loading... |
| 619 body: _body, | 629 body: _body, |
| 620 queryParams: _queryParams, | 630 queryParams: _queryParams, |
| 621 uploadOptions: _uploadOptions, | 631 uploadOptions: _uploadOptions, |
| 622 uploadMedia: _uploadMedia, | 632 uploadMedia: _uploadMedia, |
| 623 downloadOptions: _downloadOptions); | 633 downloadOptions: _downloadOptions); |
| 624 return _response.then((data) => new Job.fromJson(data)); | 634 return _response.then((data) => new Job.fromJson(data)); |
| 625 } | 635 } |
| 626 | 636 |
| 627 /** | 637 /** |
| 628 * Deletes the job from the project. If the job is active, the delete fails, | 638 * Deletes the job from the project. If the job is active, the delete fails, |
| 629 * and the response returns `FAILED_PRECONDITION`. | 639 * and the response returns FAILED_PRECONDITION. |
| 630 * | 640 * |
| 631 * Request parameters: | 641 * Request parameters: |
| 632 * | 642 * |
| 633 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 643 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 634 * the job belongs to. | 644 * job belongs to. |
| 635 * | 645 * |
| 636 * [jobId] - [Required] The job ID. | 646 * [jobId] - Required The job ID. |
| 637 * | 647 * |
| 638 * Completes with a [Empty]. | 648 * Completes with a [Empty]. |
| 639 * | 649 * |
| 640 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 650 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 641 * error. | 651 * error. |
| 642 * | 652 * |
| 643 * If the used [http.Client] completes with an error when making a REST call, | 653 * If the used [http.Client] completes with an error when making a REST call, |
| 644 * this method will complete with the same error. | 654 * this method will complete with the same error. |
| 645 */ | 655 */ |
| 646 async.Future<Empty> delete(core.String projectId, core.String jobId) { | 656 async.Future<Empty> delete(core.String projectId, core.String jobId) { |
| (...skipping 21 matching lines...) Expand all Loading... |
| 668 uploadMedia: _uploadMedia, | 678 uploadMedia: _uploadMedia, |
| 669 downloadOptions: _downloadOptions); | 679 downloadOptions: _downloadOptions); |
| 670 return _response.then((data) => new Empty.fromJson(data)); | 680 return _response.then((data) => new Empty.fromJson(data)); |
| 671 } | 681 } |
| 672 | 682 |
| 673 /** | 683 /** |
| 674 * Gets the resource representation for a job in a project. | 684 * Gets the resource representation for a job in a project. |
| 675 * | 685 * |
| 676 * Request parameters: | 686 * Request parameters: |
| 677 * | 687 * |
| 678 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 688 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 679 * the job belongs to. | 689 * job belongs to. |
| 680 * | 690 * |
| 681 * [jobId] - [Required] The job ID. | 691 * [jobId] - Required The job ID. |
| 682 * | 692 * |
| 683 * Completes with a [Job]. | 693 * Completes with a [Job]. |
| 684 * | 694 * |
| 685 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 695 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 686 * error. | 696 * error. |
| 687 * | 697 * |
| 688 * If the used [http.Client] completes with an error when making a REST call, | 698 * If the used [http.Client] completes with an error when making a REST call, |
| 689 * this method will complete with the same error. | 699 * this method will complete with the same error. |
| 690 */ | 700 */ |
| 691 async.Future<Job> get(core.String projectId, core.String jobId) { | 701 async.Future<Job> get(core.String projectId, core.String jobId) { |
| (...skipping 21 matching lines...) Expand all Loading... |
| 713 uploadMedia: _uploadMedia, | 723 uploadMedia: _uploadMedia, |
| 714 downloadOptions: _downloadOptions); | 724 downloadOptions: _downloadOptions); |
| 715 return _response.then((data) => new Job.fromJson(data)); | 725 return _response.then((data) => new Job.fromJson(data)); |
| 716 } | 726 } |
| 717 | 727 |
| 718 /** | 728 /** |
| 719 * Lists jobs in a project. | 729 * Lists jobs in a project. |
| 720 * | 730 * |
| 721 * Request parameters: | 731 * Request parameters: |
| 722 * | 732 * |
| 723 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 733 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 724 * the job belongs to. | 734 * job belongs to. |
| 725 * | 735 * |
| 726 * [pageSize] - [Optional] The number of results to return in each response. | 736 * [pageToken] - Optional The page token, returned by a previous call, to |
| 727 * | |
| 728 * [pageToken] - [Optional] The page token, returned by a previous call, to | |
| 729 * request the next page of results. | 737 * request the next page of results. |
| 730 * | 738 * |
| 731 * [clusterName] - [Optional] If set, the returned jobs list includes only | 739 * [pageSize] - Optional The number of results to return in each response. |
| 732 * jobs that were submitted to the named cluster. | |
| 733 * | 740 * |
| 734 * [jobStateMatcher] - [Optional] Specifies enumerated categories of jobs to | 741 * [clusterName] - Optional If set, the returned jobs list includes only jobs |
| 742 * that were submitted to the named cluster. |
| 743 * |
| 744 * [filter] - Optional A filter constraining which jobs to list. Valid filters |
| 745 * contain job state and label terms such as: labels.key1 = val1 AND |
| 746 * (labels.k2 = val2 OR labels.k3 = val3) |
| 747 * |
| 748 * [jobStateMatcher] - Optional Specifies enumerated categories of jobs to |
| 735 * list. | 749 * list. |
| 736 * Possible string values are: | 750 * Possible string values are: |
| 737 * - "ALL" : A ALL. | 751 * - "ALL" : A ALL. |
| 738 * - "ACTIVE" : A ACTIVE. | 752 * - "ACTIVE" : A ACTIVE. |
| 739 * - "NON_ACTIVE" : A NON_ACTIVE. | 753 * - "NON_ACTIVE" : A NON_ACTIVE. |
| 740 * | 754 * |
| 741 * [filter] - [Optional] A filter constraining which jobs to list. Valid | |
| 742 * filters contain job state and label terms such as: labels.key1 = val1 AND | |
| 743 * (labels.k2 = val2 OR labels.k3 = val3) | |
| 744 * | |
| 745 * Completes with a [ListJobsResponse]. | 755 * Completes with a [ListJobsResponse]. |
| 746 * | 756 * |
| 747 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 757 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 748 * error. | 758 * error. |
| 749 * | 759 * |
| 750 * If the used [http.Client] completes with an error when making a REST call, | 760 * If the used [http.Client] completes with an error when making a REST call, |
| 751 * this method will complete with the same error. | 761 * this method will complete with the same error. |
| 752 */ | 762 */ |
| 753 async.Future<ListJobsResponse> list(core.String projectId, {core.int pageSize,
core.String pageToken, core.String clusterName, core.String jobStateMatcher, co
re.String filter}) { | 763 async.Future<ListJobsResponse> list(core.String projectId, {core.String pageTo
ken, core.int pageSize, core.String clusterName, core.String filter, core.String
jobStateMatcher}) { |
| 754 var _url = null; | 764 var _url = null; |
| 755 var _queryParams = new core.Map(); | 765 var _queryParams = new core.Map(); |
| 756 var _uploadMedia = null; | 766 var _uploadMedia = null; |
| 757 var _uploadOptions = null; | 767 var _uploadOptions = null; |
| 758 var _downloadOptions = commons.DownloadOptions.Metadata; | 768 var _downloadOptions = commons.DownloadOptions.Metadata; |
| 759 var _body = null; | 769 var _body = null; |
| 760 | 770 |
| 761 if (projectId == null) { | 771 if (projectId == null) { |
| 762 throw new core.ArgumentError("Parameter projectId is required."); | 772 throw new core.ArgumentError("Parameter projectId is required."); |
| 763 } | 773 } |
| 774 if (pageToken != null) { |
| 775 _queryParams["pageToken"] = [pageToken]; |
| 776 } |
| 764 if (pageSize != null) { | 777 if (pageSize != null) { |
| 765 _queryParams["pageSize"] = ["${pageSize}"]; | 778 _queryParams["pageSize"] = ["${pageSize}"]; |
| 766 } | 779 } |
| 767 if (pageToken != null) { | |
| 768 _queryParams["pageToken"] = [pageToken]; | |
| 769 } | |
| 770 if (clusterName != null) { | 780 if (clusterName != null) { |
| 771 _queryParams["clusterName"] = [clusterName]; | 781 _queryParams["clusterName"] = [clusterName]; |
| 772 } | 782 } |
| 783 if (filter != null) { |
| 784 _queryParams["filter"] = [filter]; |
| 785 } |
| 773 if (jobStateMatcher != null) { | 786 if (jobStateMatcher != null) { |
| 774 _queryParams["jobStateMatcher"] = [jobStateMatcher]; | 787 _queryParams["jobStateMatcher"] = [jobStateMatcher]; |
| 775 } | 788 } |
| 776 if (filter != null) { | |
| 777 _queryParams["filter"] = [filter]; | |
| 778 } | |
| 779 | 789 |
| 780 _url = 'v1beta1/projects/' + commons.Escaper.ecapeVariable('$projectId') + '
/jobs'; | 790 _url = 'v1beta1/projects/' + commons.Escaper.ecapeVariable('$projectId') + '
/jobs'; |
| 781 | 791 |
| 782 var _response = _requester.request(_url, | 792 var _response = _requester.request(_url, |
| 783 "GET", | 793 "GET", |
| 784 body: _body, | 794 body: _body, |
| 785 queryParams: _queryParams, | 795 queryParams: _queryParams, |
| 786 uploadOptions: _uploadOptions, | 796 uploadOptions: _uploadOptions, |
| 787 uploadMedia: _uploadMedia, | 797 uploadMedia: _uploadMedia, |
| 788 downloadOptions: _downloadOptions); | 798 downloadOptions: _downloadOptions); |
| 789 return _response.then((data) => new ListJobsResponse.fromJson(data)); | 799 return _response.then((data) => new ListJobsResponse.fromJson(data)); |
| 790 } | 800 } |
| 791 | 801 |
| 792 /** | 802 /** |
| 803 * Updates a job in a project. |
| 804 * |
| 805 * [request] - The metadata request object. |
| 806 * |
| 807 * Request parameters: |
| 808 * |
| 809 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 810 * job belongs to. |
| 811 * |
| 812 * [jobId] - Required The job ID. |
| 813 * |
| 814 * [updateMask] - Required Specifies the path, relative to <code>Job</code>, |
| 815 * of the field to update. For example, to update the labels of a Job the |
| 816 * <code>update_mask</code> parameter would be specified as |
| 817 * <code>labels</code>, and the PATCH request body would specify the new |
| 818 * value. <strong>Note:</strong> Currently, <code>labels</code> is the only |
| 819 * field that can be updated. |
| 820 * |
| 821 * Completes with a [Job]. |
| 822 * |
| 823 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 824 * error. |
| 825 * |
| 826 * If the used [http.Client] completes with an error when making a REST call, |
| 827 * this method will complete with the same error. |
| 828 */ |
| 829 async.Future<Job> patch(Job request, core.String projectId, core.String jobId,
{core.String updateMask}) { |
| 830 var _url = null; |
| 831 var _queryParams = new core.Map(); |
| 832 var _uploadMedia = null; |
| 833 var _uploadOptions = null; |
| 834 var _downloadOptions = commons.DownloadOptions.Metadata; |
| 835 var _body = null; |
| 836 |
| 837 if (request != null) { |
| 838 _body = convert.JSON.encode((request).toJson()); |
| 839 } |
| 840 if (projectId == null) { |
| 841 throw new core.ArgumentError("Parameter projectId is required."); |
| 842 } |
| 843 if (jobId == null) { |
| 844 throw new core.ArgumentError("Parameter jobId is required."); |
| 845 } |
| 846 if (updateMask != null) { |
| 847 _queryParams["updateMask"] = [updateMask]; |
| 848 } |
| 849 |
| 850 _url = 'v1beta1/projects/' + commons.Escaper.ecapeVariable('$projectId') + '
/jobs/' + commons.Escaper.ecapeVariable('$jobId'); |
| 851 |
| 852 var _response = _requester.request(_url, |
| 853 "PATCH", |
| 854 body: _body, |
| 855 queryParams: _queryParams, |
| 856 uploadOptions: _uploadOptions, |
| 857 uploadMedia: _uploadMedia, |
| 858 downloadOptions: _downloadOptions); |
| 859 return _response.then((data) => new Job.fromJson(data)); |
| 860 } |
| 861 |
| 862 /** |
| 793 * Submits a job to a cluster. | 863 * Submits a job to a cluster. |
| 794 * | 864 * |
| 795 * [request] - The metadata request object. | 865 * [request] - The metadata request object. |
| 796 * | 866 * |
| 797 * Request parameters: | 867 * Request parameters: |
| 798 * | 868 * |
| 799 * [projectId] - [Required] The ID of the Google Cloud Platform project that | 869 * [projectId] - Required The ID of the Google Cloud Platform project that the |
| 800 * the job belongs to. | 870 * job belongs to. |
| 801 * | 871 * |
| 802 * Completes with a [Job]. | 872 * Completes with a [Job]. |
| 803 * | 873 * |
| 804 * Completes with a [commons.ApiRequestError] if the API endpoint returned an | 874 * Completes with a [commons.ApiRequestError] if the API endpoint returned an |
| 805 * error. | 875 * error. |
| 806 * | 876 * |
| 807 * If the used [http.Client] completes with an error when making a REST call, | 877 * If the used [http.Client] completes with an error when making a REST call, |
| 808 * this method will complete with the same error. | 878 * this method will complete with the same error. |
| 809 */ | 879 */ |
| 810 async.Future<Job> submit(SubmitJobRequest request, core.String projectId) { | 880 async.Future<Job> submit(SubmitJobRequest request, core.String projectId) { |
| (...skipping 54 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 865 return _json; | 935 return _json; |
| 866 } | 936 } |
| 867 } | 937 } |
| 868 | 938 |
| 869 /** | 939 /** |
| 870 * Describes the identifying information, configuration, and status of a cluster | 940 * Describes the identifying information, configuration, and status of a cluster |
| 871 * of Google Compute Engine instances. | 941 * of Google Compute Engine instances. |
| 872 */ | 942 */ |
| 873 class Cluster { | 943 class Cluster { |
| 874 /** | 944 /** |
| 875 * [Required] The cluster name. Cluster names within a project must be unique. | 945 * Required The cluster name. Cluster names within a project must be unique. |
| 876 * Names from deleted clusters can be reused. | 946 * Names from deleted clusters can be reused. |
| 877 */ | 947 */ |
| 878 core.String clusterName; | 948 core.String clusterName; |
| 879 /** | 949 /** |
| 880 * [Output-only] A cluster UUID (Unique Universal Identifier). Cloud Dataproc | 950 * Output-only A cluster UUID (Unique Universal Identifier). Cloud Dataproc |
| 881 * generates this value when it creates the cluster. | 951 * generates this value when it creates the cluster. |
| 882 */ | 952 */ |
| 883 core.String clusterUuid; | 953 core.String clusterUuid; |
| 884 /** | 954 /** |
| 885 * [Required] The cluster configuration. Note that Cloud Dataproc may set | 955 * Required The cluster configuration. Note that Cloud Dataproc may set |
| 886 * default values, and values may change when clusters are updated. | 956 * default values, and values may change when clusters are updated. |
| 887 */ | 957 */ |
| 888 ClusterConfiguration configuration; | 958 ClusterConfiguration configuration; |
| 889 /** | 959 /** |
| 890 * [Optional] The labels to associate with this cluster. Label keys must be | 960 * Optional The labels to associate with this cluster.Label keys must be |
| 891 * between 1 and 63 characters long, and must conform to the following PCRE | 961 * between 1 and 63 characters long, and must conform to the following PCRE |
| 892 * regular expression: \p{Ll}\p{Lo}{0,62} Label values must be between 1 and | 962 * regular expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 |
| 893 * 63 characters long, and must conform to the following PCRE regular | 963 * characters long, and must conform to the following PCRE regular expression: |
| 894 * expression: [\p{Ll}\p{Lo}\p{N}_-]{0,63} No more than 64 labels can be | 964 * \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 64 labels can be associated with a |
| 895 * associated with a given cluster. | 965 * given cluster. |
| 896 */ | 966 */ |
| 897 core.Map<core.String, core.String> labels; | 967 core.Map<core.String, core.String> labels; |
| 898 /** Contains cluster daemon metrics such as HDFS and YARN stats. */ | 968 /** Contains cluster daemon metrics such as HDFS and YARN stats. */ |
| 899 ClusterMetrics metrics; | 969 ClusterMetrics metrics; |
| 900 /** | 970 /** |
| 901 * [Required] The Google Cloud Platform project ID that the cluster belongs | 971 * Required The Google Cloud Platform project ID that the cluster belongs to. |
| 902 * to. | |
| 903 */ | 972 */ |
| 904 core.String projectId; | 973 core.String projectId; |
| 905 /** [Output-only] Cluster status. */ | 974 /** Output-only Cluster status. */ |
| 906 ClusterStatus status; | 975 ClusterStatus status; |
| 907 /** [Output-only] Previous cluster statuses. */ | 976 /** Output-only Previous cluster statuses. */ |
| 908 core.List<ClusterStatus> statusHistory; | 977 core.List<ClusterStatus> statusHistory; |
| 909 | 978 |
| 910 Cluster(); | 979 Cluster(); |
| 911 | 980 |
| 912 Cluster.fromJson(core.Map _json) { | 981 Cluster.fromJson(core.Map _json) { |
| 913 if (_json.containsKey("clusterName")) { | 982 if (_json.containsKey("clusterName")) { |
| 914 clusterName = _json["clusterName"]; | 983 clusterName = _json["clusterName"]; |
| 915 } | 984 } |
| 916 if (_json.containsKey("clusterUuid")) { | 985 if (_json.containsKey("clusterUuid")) { |
| 917 clusterUuid = _json["clusterUuid"]; | 986 clusterUuid = _json["clusterUuid"]; |
| (...skipping 44 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 962 if (statusHistory != null) { | 1031 if (statusHistory != null) { |
| 963 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); | 1032 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); |
| 964 } | 1033 } |
| 965 return _json; | 1034 return _json; |
| 966 } | 1035 } |
| 967 } | 1036 } |
| 968 | 1037 |
| 969 /** The cluster configuration. */ | 1038 /** The cluster configuration. */ |
| 970 class ClusterConfiguration { | 1039 class ClusterConfiguration { |
| 971 /** | 1040 /** |
| 972 * [Optional] A Google Cloud Storage staging bucket used for sharing generated | 1041 * Optional A Google Cloud Storage staging bucket used for sharing generated |
| 973 * SSH keys and configuration. If you do not specify a staging bucket, Cloud | 1042 * SSH keys and configuration. If you do not specify a staging bucket, Cloud |
| 974 * Dataproc will determine an appropriate Cloud Storage location (US, ASIA, or | 1043 * Dataproc will determine an appropriate Cloud Storage location (US, ASIA, or |
| 975 * EU) for your cluster's staging bucket according to the Google Compute | 1044 * EU) for your cluster's staging bucket according to the Google Compute |
| 976 * Engine zone where your cluster is deployed, and then it will create and | 1045 * Engine zone where your cluster is deployed, and then it will create and |
| 977 * manage this project-level, per-location bucket for you. | 1046 * manage this project-level, per-location bucket for you. |
| 978 */ | 1047 */ |
| 979 core.String configurationBucket; | 1048 core.String configurationBucket; |
| 980 /** | 1049 /** |
| 981 * [Required] The shared Google Compute Engine configuration settings for all | 1050 * Required The shared Google Compute Engine configuration settings for all |
| 982 * instances in a cluster. | 1051 * instances in a cluster. |
| 983 */ | 1052 */ |
| 984 GceClusterConfiguration gceClusterConfiguration; | 1053 GceClusterConfiguration gceClusterConfiguration; |
| 985 /** | 1054 /** |
| 986 * [Optional] Commands to execute on each node after configuration is | 1055 * Optional Commands to execute on each node after configuration is completed. |
| 987 * completed. By default, executables are run on master and all worker nodes. | 1056 * By default, executables are run on master and all worker nodes. You can |
| 988 * You can test a node's role metadata to run an executable on a master or | 1057 * test a node's <code>role</code> metadata to run an executable on a master |
| 989 * worker node, as shown below: ROLE=$(/usr/share/google/get_metadata_value | 1058 * or worker node, as shown below: |
| 990 * attributes/role) if [[ "${ROLE}" == 'Master' ]]; then ... master specific | 1059 * ROLE=$(/usr/share/google/get_metadata_value attributes/role) |
| 991 * actions ... else ... worker specific actions ... fi | 1060 * if [[ "${ROLE}" == 'Master' ]]; then |
| 1061 * ... master specific actions ... |
| 1062 * else |
| 1063 * ... worker specific actions ... |
| 1064 * fi |
| 992 */ | 1065 */ |
| 993 core.List<NodeInitializationAction> initializationActions; | 1066 core.List<NodeInitializationAction> initializationActions; |
| 994 /** | 1067 /** |
| 995 * [Optional] The Google Compute Engine configuration settings for the master | 1068 * Optional The Google Compute Engine configuration settings for the master |
| 996 * instance in a cluster. | 1069 * instance in a cluster. |
| 997 */ | 1070 */ |
| 998 InstanceGroupConfiguration masterConfiguration; | 1071 InstanceGroupConfiguration masterConfiguration; |
| 999 /** | 1072 /** |
| 1000 * [Optional] The Google Compute Engine configuration settings for additional | 1073 * Optional The Google Compute Engine configuration settings for additional |
| 1001 * worker instances in a cluster. | 1074 * worker instances in a cluster. |
| 1002 */ | 1075 */ |
| 1003 InstanceGroupConfiguration secondaryWorkerConfiguration; | 1076 InstanceGroupConfiguration secondaryWorkerConfiguration; |
| 1004 /** [Optional] The configuration settings for software inside the cluster. */ | 1077 /** Optional The configuration settings for software inside the cluster. */ |
| 1005 SoftwareConfiguration softwareConfiguration; | 1078 SoftwareConfiguration softwareConfiguration; |
| 1006 /** | 1079 /** |
| 1007 * [Optional] The Google Compute Engine configuration settings for worker | 1080 * Optional The Google Compute Engine configuration settings for worker |
| 1008 * instances in a cluster. | 1081 * instances in a cluster. |
| 1009 */ | 1082 */ |
| 1010 InstanceGroupConfiguration workerConfiguration; | 1083 InstanceGroupConfiguration workerConfiguration; |
| 1011 | 1084 |
| 1012 ClusterConfiguration(); | 1085 ClusterConfiguration(); |
| 1013 | 1086 |
| 1014 ClusterConfiguration.fromJson(core.Map _json) { | 1087 ClusterConfiguration.fromJson(core.Map _json) { |
| 1015 if (_json.containsKey("configurationBucket")) { | 1088 if (_json.containsKey("configurationBucket")) { |
| 1016 configurationBucket = _json["configurationBucket"]; | 1089 configurationBucket = _json["configurationBucket"]; |
| 1017 } | 1090 } |
| (...skipping 69 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1087 } | 1160 } |
| 1088 if (yarnMetrics != null) { | 1161 if (yarnMetrics != null) { |
| 1089 _json["yarnMetrics"] = yarnMetrics; | 1162 _json["yarnMetrics"] = yarnMetrics; |
| 1090 } | 1163 } |
| 1091 return _json; | 1164 return _json; |
| 1092 } | 1165 } |
| 1093 } | 1166 } |
| 1094 | 1167 |
| 1095 /** Metadata describing the operation. */ | 1168 /** Metadata describing the operation. */ |
| 1096 class ClusterOperationMetadata { | 1169 class ClusterOperationMetadata { |
| 1097 /** [Output-only] Name of the cluster for the operation. */ | 1170 /** Output-only Name of the cluster for the operation. */ |
| 1098 core.String clusterName; | 1171 core.String clusterName; |
| 1099 /** [Output-only] Cluster UUID for the operation. */ | 1172 /** Output-only Cluster UUID for the operation. */ |
| 1100 core.String clusterUuid; | 1173 core.String clusterUuid; |
| 1101 /** [Output-only] Short description of operation. */ | 1174 /** Output-only Short description of operation. */ |
| 1102 core.String description; | 1175 core.String description; |
| 1103 /** [Output-only] labels associated with the operation */ | 1176 /** Output-only Labels associated with the operation */ |
| 1104 core.Map<core.String, core.String> labels; | 1177 core.Map<core.String, core.String> labels; |
| 1105 /** [Output-only] The operation type. */ | 1178 /** Output-only The operation type. */ |
| 1106 core.String operationType; | 1179 core.String operationType; |
| 1107 /** [Output-only] Current operation status. */ | 1180 /** Output-only Current operation status. */ |
| 1108 ClusterOperationStatus status; | 1181 ClusterOperationStatus status; |
| 1109 /** [Output-only] The previous operation status. */ | 1182 /** Output-only The previous operation status. */ |
| 1110 core.List<ClusterOperationStatus> statusHistory; | 1183 core.List<ClusterOperationStatus> statusHistory; |
| 1184 /** Output-only Errors encountered during operation execution. */ |
| 1185 core.List<core.String> warnings; |
| 1111 | 1186 |
| 1112 ClusterOperationMetadata(); | 1187 ClusterOperationMetadata(); |
| 1113 | 1188 |
| 1114 ClusterOperationMetadata.fromJson(core.Map _json) { | 1189 ClusterOperationMetadata.fromJson(core.Map _json) { |
| 1115 if (_json.containsKey("clusterName")) { | 1190 if (_json.containsKey("clusterName")) { |
| 1116 clusterName = _json["clusterName"]; | 1191 clusterName = _json["clusterName"]; |
| 1117 } | 1192 } |
| 1118 if (_json.containsKey("clusterUuid")) { | 1193 if (_json.containsKey("clusterUuid")) { |
| 1119 clusterUuid = _json["clusterUuid"]; | 1194 clusterUuid = _json["clusterUuid"]; |
| 1120 } | 1195 } |
| 1121 if (_json.containsKey("description")) { | 1196 if (_json.containsKey("description")) { |
| 1122 description = _json["description"]; | 1197 description = _json["description"]; |
| 1123 } | 1198 } |
| 1124 if (_json.containsKey("labels")) { | 1199 if (_json.containsKey("labels")) { |
| 1125 labels = _json["labels"]; | 1200 labels = _json["labels"]; |
| 1126 } | 1201 } |
| 1127 if (_json.containsKey("operationType")) { | 1202 if (_json.containsKey("operationType")) { |
| 1128 operationType = _json["operationType"]; | 1203 operationType = _json["operationType"]; |
| 1129 } | 1204 } |
| 1130 if (_json.containsKey("status")) { | 1205 if (_json.containsKey("status")) { |
| 1131 status = new ClusterOperationStatus.fromJson(_json["status"]); | 1206 status = new ClusterOperationStatus.fromJson(_json["status"]); |
| 1132 } | 1207 } |
| 1133 if (_json.containsKey("statusHistory")) { | 1208 if (_json.containsKey("statusHistory")) { |
| 1134 statusHistory = _json["statusHistory"].map((value) => new ClusterOperation
Status.fromJson(value)).toList(); | 1209 statusHistory = _json["statusHistory"].map((value) => new ClusterOperation
Status.fromJson(value)).toList(); |
| 1135 } | 1210 } |
| 1211 if (_json.containsKey("warnings")) { |
| 1212 warnings = _json["warnings"]; |
| 1213 } |
| 1136 } | 1214 } |
| 1137 | 1215 |
| 1138 core.Map toJson() { | 1216 core.Map toJson() { |
| 1139 var _json = new core.Map(); | 1217 var _json = new core.Map(); |
| 1140 if (clusterName != null) { | 1218 if (clusterName != null) { |
| 1141 _json["clusterName"] = clusterName; | 1219 _json["clusterName"] = clusterName; |
| 1142 } | 1220 } |
| 1143 if (clusterUuid != null) { | 1221 if (clusterUuid != null) { |
| 1144 _json["clusterUuid"] = clusterUuid; | 1222 _json["clusterUuid"] = clusterUuid; |
| 1145 } | 1223 } |
| 1146 if (description != null) { | 1224 if (description != null) { |
| 1147 _json["description"] = description; | 1225 _json["description"] = description; |
| 1148 } | 1226 } |
| 1149 if (labels != null) { | 1227 if (labels != null) { |
| 1150 _json["labels"] = labels; | 1228 _json["labels"] = labels; |
| 1151 } | 1229 } |
| 1152 if (operationType != null) { | 1230 if (operationType != null) { |
| 1153 _json["operationType"] = operationType; | 1231 _json["operationType"] = operationType; |
| 1154 } | 1232 } |
| 1155 if (status != null) { | 1233 if (status != null) { |
| 1156 _json["status"] = (status).toJson(); | 1234 _json["status"] = (status).toJson(); |
| 1157 } | 1235 } |
| 1158 if (statusHistory != null) { | 1236 if (statusHistory != null) { |
| 1159 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); | 1237 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); |
| 1160 } | 1238 } |
| 1239 if (warnings != null) { |
| 1240 _json["warnings"] = warnings; |
| 1241 } |
| 1161 return _json; | 1242 return _json; |
| 1162 } | 1243 } |
| 1163 } | 1244 } |
| 1164 | 1245 |
| 1165 /** The status of the operation. */ | 1246 /** The status of the operation. */ |
| 1166 class ClusterOperationStatus { | 1247 class ClusterOperationStatus { |
| 1167 /** [Output-only]A message containing any operation metadata details. */ | 1248 /** Output-onlyA message containing any operation metadata details. */ |
| 1168 core.String details; | 1249 core.String details; |
| 1169 /** [Output-only] A message containing the detailed operation state. */ | 1250 /** Output-only A message containing the detailed operation state. */ |
| 1170 core.String innerState; | 1251 core.String innerState; |
| 1171 /** | 1252 /** |
| 1172 * [Output-only] A message containing the operation state. | 1253 * Output-only A message containing the operation state. |
| 1173 * Possible string values are: | 1254 * Possible string values are: |
| 1174 * - "UNKNOWN" : A UNKNOWN. | 1255 * - "UNKNOWN" : Unused. |
| 1175 * - "PENDING" : A PENDING. | 1256 * - "PENDING" : The operation has been created. |
| 1176 * - "RUNNING" : A RUNNING. | 1257 * - "RUNNING" : The operation is running. |
| 1177 * - "DONE" : A DONE. | 1258 * - "DONE" : The operation is done; either cancelled or completed. |
| 1178 */ | 1259 */ |
| 1179 core.String state; | 1260 core.String state; |
| 1180 /** [Output-only] The time this state was entered. */ | 1261 /** Output-only The time this state was entered. */ |
| 1181 core.String stateStartTime; | 1262 core.String stateStartTime; |
| 1182 | 1263 |
| 1183 ClusterOperationStatus(); | 1264 ClusterOperationStatus(); |
| 1184 | 1265 |
| 1185 ClusterOperationStatus.fromJson(core.Map _json) { | 1266 ClusterOperationStatus.fromJson(core.Map _json) { |
| 1186 if (_json.containsKey("details")) { | 1267 if (_json.containsKey("details")) { |
| 1187 details = _json["details"]; | 1268 details = _json["details"]; |
| 1188 } | 1269 } |
| 1189 if (_json.containsKey("innerState")) { | 1270 if (_json.containsKey("innerState")) { |
| 1190 innerState = _json["innerState"]; | 1271 innerState = _json["innerState"]; |
| (...skipping 24 matching lines...) Expand all Loading... |
| 1215 } | 1296 } |
| 1216 } | 1297 } |
| 1217 | 1298 |
| 1218 /** The status of a cluster and its instances. */ | 1299 /** The status of a cluster and its instances. */ |
| 1219 class ClusterStatus { | 1300 class ClusterStatus { |
| 1220 /** Optional details of cluster's state. */ | 1301 /** Optional details of cluster's state. */ |
| 1221 core.String detail; | 1302 core.String detail; |
| 1222 /** | 1303 /** |
| 1223 * The cluster's state. | 1304 * The cluster's state. |
| 1224 * Possible string values are: | 1305 * Possible string values are: |
| 1225 * - "UNKNOWN" : A UNKNOWN. | 1306 * - "UNKNOWN" : The cluster state is unknown. |
| 1226 * - "CREATING" : A CREATING. | 1307 * - "CREATING" : The cluster is being created and set up. It is not ready for |
| 1227 * - "RUNNING" : A RUNNING. | 1308 * use. |
| 1228 * - "ERROR" : A ERROR. | 1309 * - "RUNNING" : The cluster is currently running and healthy. It is ready for |
| 1229 * - "DELETING" : A DELETING. | 1310 * use. |
| 1230 * - "UPDATING" : A UPDATING. | 1311 * - "ERROR" : The cluster encountered an error. It is not ready for use. |
| 1312 * - "DELETING" : The cluster is being deleted. It cannot be used. |
| 1313 * - "UPDATING" : The cluster is being updated. It continues to accept and |
| 1314 * process jobs. |
| 1231 */ | 1315 */ |
| 1232 core.String state; | 1316 core.String state; |
| 1233 /** Time when this state was entered. */ | 1317 /** Time when this state was entered. */ |
| 1234 core.String stateStartTime; | 1318 core.String stateStartTime; |
| 1235 | 1319 |
| 1236 ClusterStatus(); | 1320 ClusterStatus(); |
| 1237 | 1321 |
| 1238 ClusterStatus.fromJson(core.Map _json) { | 1322 ClusterStatus.fromJson(core.Map _json) { |
| 1239 if (_json.containsKey("detail")) { | 1323 if (_json.containsKey("detail")) { |
| 1240 detail = _json["detail"]; | 1324 detail = _json["detail"]; |
| (...skipping 17 matching lines...) Expand all Loading... |
| 1258 if (stateStartTime != null) { | 1342 if (stateStartTime != null) { |
| 1259 _json["stateStartTime"] = stateStartTime; | 1343 _json["stateStartTime"] = stateStartTime; |
| 1260 } | 1344 } |
| 1261 return _json; | 1345 return _json; |
| 1262 } | 1346 } |
| 1263 } | 1347 } |
| 1264 | 1348 |
| 1265 /** The location where output from diagnostic command can be found. */ | 1349 /** The location where output from diagnostic command can be found. */ |
| 1266 class DiagnoseClusterOutputLocation { | 1350 class DiagnoseClusterOutputLocation { |
| 1267 /** | 1351 /** |
| 1268 * [Output-only] The Google Cloud Storage URI of the diagnostic output. This | 1352 * Output-only The Google Cloud Storage URI of the diagnostic output. This |
| 1269 * will be a plain text file with summary of collected diagnostics. | 1353 * will be a plain text file with summary of collected diagnostics. |
| 1270 */ | 1354 */ |
| 1271 core.String outputUri; | 1355 core.String outputUri; |
| 1272 | 1356 |
| 1273 DiagnoseClusterOutputLocation(); | 1357 DiagnoseClusterOutputLocation(); |
| 1274 | 1358 |
| 1275 DiagnoseClusterOutputLocation.fromJson(core.Map _json) { | 1359 DiagnoseClusterOutputLocation.fromJson(core.Map _json) { |
| 1276 if (_json.containsKey("outputUri")) { | 1360 if (_json.containsKey("outputUri")) { |
| 1277 outputUri = _json["outputUri"]; | 1361 outputUri = _json["outputUri"]; |
| 1278 } | 1362 } |
| (...skipping 18 matching lines...) Expand all Loading... |
| 1297 | 1381 |
| 1298 core.Map toJson() { | 1382 core.Map toJson() { |
| 1299 var _json = new core.Map(); | 1383 var _json = new core.Map(); |
| 1300 return _json; | 1384 return _json; |
| 1301 } | 1385 } |
| 1302 } | 1386 } |
| 1303 | 1387 |
| 1304 /** The location of diagnostic output. */ | 1388 /** The location of diagnostic output. */ |
| 1305 class DiagnoseClusterResults { | 1389 class DiagnoseClusterResults { |
| 1306 /** | 1390 /** |
| 1307 * [Output-only] The Google Cloud Storage URI of the diagnostic output. The | 1391 * Output-only The Google Cloud Storage URI of the diagnostic output. The |
| 1308 * output report is a plain text file with a summary of collected diagnostics. | 1392 * output report is a plain text file with a summary of collected diagnostics. |
| 1309 */ | 1393 */ |
| 1310 core.String outputUri; | 1394 core.String outputUri; |
| 1311 | 1395 |
| 1312 DiagnoseClusterResults(); | 1396 DiagnoseClusterResults(); |
| 1313 | 1397 |
| 1314 DiagnoseClusterResults.fromJson(core.Map _json) { | 1398 DiagnoseClusterResults.fromJson(core.Map _json) { |
| 1315 if (_json.containsKey("outputUri")) { | 1399 if (_json.containsKey("outputUri")) { |
| 1316 outputUri = _json["outputUri"]; | 1400 outputUri = _json["outputUri"]; |
| 1317 } | 1401 } |
| 1318 } | 1402 } |
| 1319 | 1403 |
| 1320 core.Map toJson() { | 1404 core.Map toJson() { |
| 1321 var _json = new core.Map(); | 1405 var _json = new core.Map(); |
| 1322 if (outputUri != null) { | 1406 if (outputUri != null) { |
| 1323 _json["outputUri"] = outputUri; | 1407 _json["outputUri"] = outputUri; |
| 1324 } | 1408 } |
| 1325 return _json; | 1409 return _json; |
| 1326 } | 1410 } |
| 1327 } | 1411 } |
| 1328 | 1412 |
| 1329 /** Specifies the configuration of disk options for a group of VM instances. */ | 1413 /** Specifies the configuration of disk options for a group of VM instances. */ |
| 1330 class DiskConfiguration { | 1414 class DiskConfiguration { |
| 1331 /** [Optional] Size in GB of the boot disk (default is 500GB). */ | 1415 /** Optional Size in GB of the boot disk (default is 500GB). */ |
| 1332 core.int bootDiskSizeGb; | 1416 core.int bootDiskSizeGb; |
| 1333 /** | 1417 /** |
| 1334 * [Optional] Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are | 1418 * Optional Number of attached SSDs, from 0 to 4 (default is 0). If SSDs are |
| 1335 * not attached, the boot disk is used to store runtime logs and HDFS data. If | 1419 * not attached, the boot disk is used to store runtime logs and HDFS data. If |
| 1336 * one or more SSDs are attached, this runtime bulk data is spread across | 1420 * one or more SSDs are attached, this runtime bulk data is spread across |
| 1337 * them, and the boot disk contains only basic configuration and installed | 1421 * them, and the boot disk contains only basic configuration and installed |
| 1338 * binaries. | 1422 * binaries. |
| 1339 */ | 1423 */ |
| 1340 core.int numLocalSsds; | 1424 core.int numLocalSsds; |
| 1341 | 1425 |
| 1342 DiskConfiguration(); | 1426 DiskConfiguration(); |
| 1343 | 1427 |
| 1344 DiskConfiguration.fromJson(core.Map _json) { | 1428 DiskConfiguration.fromJson(core.Map _json) { |
| (...skipping 13 matching lines...) Expand all Loading... |
| 1358 if (numLocalSsds != null) { | 1442 if (numLocalSsds != null) { |
| 1359 _json["numLocalSsds"] = numLocalSsds; | 1443 _json["numLocalSsds"] = numLocalSsds; |
| 1360 } | 1444 } |
| 1361 return _json; | 1445 return _json; |
| 1362 } | 1446 } |
| 1363 } | 1447 } |
| 1364 | 1448 |
| 1365 /** | 1449 /** |
| 1366 * A generic empty message that you can re-use to avoid defining duplicated | 1450 * A generic empty message that you can re-use to avoid defining duplicated |
| 1367 * empty messages in your APIs. A typical example is to use it as the request or | 1451 * empty messages in your APIs. A typical example is to use it as the request or |
| 1368 * the response type of an API method. For instance: service Foo { rpc | 1452 * the response type of an API method. For instance: |
| 1369 * Bar(google.protobuf.Empty) returns (google.protobuf.Empty); } The JSON | 1453 * service Foo { |
| 1370 * representation for `Empty` is empty JSON object `{}`. | 1454 * rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); |
| 1455 * } |
| 1456 * The JSON representation for Empty is empty JSON object {}. |
| 1371 */ | 1457 */ |
| 1372 class Empty { | 1458 class Empty { |
| 1373 | 1459 |
| 1374 Empty(); | 1460 Empty(); |
| 1375 | 1461 |
| 1376 Empty.fromJson(core.Map _json) { | 1462 Empty.fromJson(core.Map _json) { |
| 1377 } | 1463 } |
| 1378 | 1464 |
| 1379 core.Map toJson() { | 1465 core.Map toJson() { |
| 1380 var _json = new core.Map(); | 1466 var _json = new core.Map(); |
| (...skipping 17 matching lines...) Expand all Loading... |
| 1398 core.bool internalIpOnly; | 1484 core.bool internalIpOnly; |
| 1399 /** The Google Compute Engine metadata entries to add to all instances. */ | 1485 /** The Google Compute Engine metadata entries to add to all instances. */ |
| 1400 core.Map<core.String, core.String> metadata; | 1486 core.Map<core.String, core.String> metadata; |
| 1401 /** | 1487 /** |
| 1402 * The Google Compute Engine network to be used for machine communications. | 1488 * The Google Compute Engine network to be used for machine communications. |
| 1403 * Cannot be specified with subnetwork_uri. If neither network_uri nor | 1489 * Cannot be specified with subnetwork_uri. If neither network_uri nor |
| 1404 * subnetwork_uri is specified, the "default" network of the project is used, | 1490 * subnetwork_uri is specified, the "default" network of the project is used, |
| 1405 * if it exists. Cannot be a "Custom Subnet Network" (see | 1491 * if it exists. Cannot be a "Custom Subnet Network" (see |
| 1406 * https://cloud.google.com/compute/docs/subnetworks for more information). | 1492 * https://cloud.google.com/compute/docs/subnetworks for more information). |
| 1407 * Example: | 1493 * Example: |
| 1408 * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global
/default`. | 1494 * https://www.googleapis.com/compute/v1/projects/[project_id]/regions/global/
default. |
| 1409 */ | 1495 */ |
| 1410 core.String networkUri; | 1496 core.String networkUri; |
| 1411 /** | 1497 /** |
| 1498 * Optional The service account of the instances. Defaults to the default |
| 1499 * Google Compute Engine service account. Custom service accounts need |
| 1500 * permissions equivalent to the folloing IAM roles: |
| 1501 * roles/logging.logWriter |
| 1502 * roles/storage.objectAdmin(see |
| 1503 * https://cloud.google.com/compute/docs/access/service-accounts#custom_servic
e_accounts |
| 1504 * for more information). Example: |
| 1505 * [account_id]@[project_id].iam.gserviceaccount.com |
| 1506 */ |
| 1507 core.String serviceAccount; |
| 1508 /** |
| 1412 * The URIs of service account scopes to be included in Google Compute Engine | 1509 * The URIs of service account scopes to be included in Google Compute Engine |
| 1413 * instances. The following base set of scopes is always included: - | 1510 * instances. The following base set of scopes is always included: - |
| 1414 * https://www.googleapis.com/auth/cloud.useraccounts.readonly - | 1511 * https://www.googleapis.com/auth/cloud.useraccounts.readonly - |
| 1415 * https://www.googleapis.com/auth/devstorage.read_write - | 1512 * https://www.googleapis.com/auth/devstorage.read_write - |
| 1416 * https://www.googleapis.com/auth/logging.write If no scopes are specfied, | 1513 * https://www.googleapis.com/auth/logging.write If no scopes are specfied, |
| 1417 * the following defaults are also provided: - | 1514 * the following defaults are also provided: - |
| 1418 * https://www.googleapis.com/auth/bigquery - | 1515 * https://www.googleapis.com/auth/bigquery - |
| 1419 * https://www.googleapis.com/auth/bigtable.admin.table - | 1516 * https://www.googleapis.com/auth/bigtable.admin.table - |
| 1420 * https://www.googleapis.com/auth/bigtable.data - | 1517 * https://www.googleapis.com/auth/bigtable.data - |
| 1421 * https://www.googleapis.com/auth/devstorage.full_control | 1518 * https://www.googleapis.com/auth/devstorage.full_control |
| 1422 */ | 1519 */ |
| 1423 core.List<core.String> serviceAccountScopes; | 1520 core.List<core.String> serviceAccountScopes; |
| 1424 /** | 1521 /** |
| 1425 * The Google Compute Engine subnetwork to be used for machine communications. | 1522 * The Google Compute Engine subnetwork to be used for machine communications. |
| 1426 * Cannot be specified with network_uri. Example: | 1523 * Cannot be specified with network_uri. Example: |
| 1427 * `https://www.googleapis.com/compute/v1/projects/[project_id]/regions/us-eas
t1/sub0`. | 1524 * https://www.googleapis.com/compute/v1/projects/[project_id]/regions/us-east
1/sub0. |
| 1428 */ | 1525 */ |
| 1429 core.String subnetworkUri; | 1526 core.String subnetworkUri; |
| 1430 /** The Google Compute Engine tags to add to all instances. */ | 1527 /** The Google Compute Engine tags to add to all instances. */ |
| 1431 core.List<core.String> tags; | 1528 core.List<core.String> tags; |
| 1432 /** | 1529 /** |
| 1433 * [Required] The zone where the Google Compute Engine cluster will be | 1530 * Required The zone where the Google Compute Engine cluster will be located. |
| 1434 * located. Example: | 1531 * Example: |
| 1435 * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]`. | 1532 * https://www.googleapis.com/compute/v1/projects/[project_id]/zones/[zone]. |
| 1436 */ | 1533 */ |
| 1437 core.String zoneUri; | 1534 core.String zoneUri; |
| 1438 | 1535 |
| 1439 GceClusterConfiguration(); | 1536 GceClusterConfiguration(); |
| 1440 | 1537 |
| 1441 GceClusterConfiguration.fromJson(core.Map _json) { | 1538 GceClusterConfiguration.fromJson(core.Map _json) { |
| 1442 if (_json.containsKey("internalIpOnly")) { | 1539 if (_json.containsKey("internalIpOnly")) { |
| 1443 internalIpOnly = _json["internalIpOnly"]; | 1540 internalIpOnly = _json["internalIpOnly"]; |
| 1444 } | 1541 } |
| 1445 if (_json.containsKey("metadata")) { | 1542 if (_json.containsKey("metadata")) { |
| 1446 metadata = _json["metadata"]; | 1543 metadata = _json["metadata"]; |
| 1447 } | 1544 } |
| 1448 if (_json.containsKey("networkUri")) { | 1545 if (_json.containsKey("networkUri")) { |
| 1449 networkUri = _json["networkUri"]; | 1546 networkUri = _json["networkUri"]; |
| 1450 } | 1547 } |
| 1548 if (_json.containsKey("serviceAccount")) { |
| 1549 serviceAccount = _json["serviceAccount"]; |
| 1550 } |
| 1451 if (_json.containsKey("serviceAccountScopes")) { | 1551 if (_json.containsKey("serviceAccountScopes")) { |
| 1452 serviceAccountScopes = _json["serviceAccountScopes"]; | 1552 serviceAccountScopes = _json["serviceAccountScopes"]; |
| 1453 } | 1553 } |
| 1454 if (_json.containsKey("subnetworkUri")) { | 1554 if (_json.containsKey("subnetworkUri")) { |
| 1455 subnetworkUri = _json["subnetworkUri"]; | 1555 subnetworkUri = _json["subnetworkUri"]; |
| 1456 } | 1556 } |
| 1457 if (_json.containsKey("tags")) { | 1557 if (_json.containsKey("tags")) { |
| 1458 tags = _json["tags"]; | 1558 tags = _json["tags"]; |
| 1459 } | 1559 } |
| 1460 if (_json.containsKey("zoneUri")) { | 1560 if (_json.containsKey("zoneUri")) { |
| 1461 zoneUri = _json["zoneUri"]; | 1561 zoneUri = _json["zoneUri"]; |
| 1462 } | 1562 } |
| 1463 } | 1563 } |
| 1464 | 1564 |
| 1465 core.Map toJson() { | 1565 core.Map toJson() { |
| 1466 var _json = new core.Map(); | 1566 var _json = new core.Map(); |
| 1467 if (internalIpOnly != null) { | 1567 if (internalIpOnly != null) { |
| 1468 _json["internalIpOnly"] = internalIpOnly; | 1568 _json["internalIpOnly"] = internalIpOnly; |
| 1469 } | 1569 } |
| 1470 if (metadata != null) { | 1570 if (metadata != null) { |
| 1471 _json["metadata"] = metadata; | 1571 _json["metadata"] = metadata; |
| 1472 } | 1572 } |
| 1473 if (networkUri != null) { | 1573 if (networkUri != null) { |
| 1474 _json["networkUri"] = networkUri; | 1574 _json["networkUri"] = networkUri; |
| 1475 } | 1575 } |
| 1576 if (serviceAccount != null) { |
| 1577 _json["serviceAccount"] = serviceAccount; |
| 1578 } |
| 1476 if (serviceAccountScopes != null) { | 1579 if (serviceAccountScopes != null) { |
| 1477 _json["serviceAccountScopes"] = serviceAccountScopes; | 1580 _json["serviceAccountScopes"] = serviceAccountScopes; |
| 1478 } | 1581 } |
| 1479 if (subnetworkUri != null) { | 1582 if (subnetworkUri != null) { |
| 1480 _json["subnetworkUri"] = subnetworkUri; | 1583 _json["subnetworkUri"] = subnetworkUri; |
| 1481 } | 1584 } |
| 1482 if (tags != null) { | 1585 if (tags != null) { |
| 1483 _json["tags"] = tags; | 1586 _json["tags"] = tags; |
| 1484 } | 1587 } |
| 1485 if (zoneUri != null) { | 1588 if (zoneUri != null) { |
| 1486 _json["zoneUri"] = zoneUri; | 1589 _json["zoneUri"] = zoneUri; |
| 1487 } | 1590 } |
| 1488 return _json; | 1591 return _json; |
| 1489 } | 1592 } |
| 1490 } | 1593 } |
| 1491 | 1594 |
| 1492 /** A Cloud Dataproc job for running Hadoop MapReduce jobs on YARN. */ | 1595 /** A Cloud Dataproc job for running Hadoop MapReduce jobs on YARN. */ |
| 1493 class HadoopJob { | 1596 class HadoopJob { |
| 1494 /** | 1597 /** |
| 1495 * [Optional] HCFS URIs of archives to be extracted in the working directory | 1598 * Optional HCFS URIs of archives to be extracted in the working directory of |
| 1496 * of Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, | 1599 * Hadoop drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, |
| 1497 * .tgz, or .zip. | 1600 * or .zip. |
| 1498 */ | 1601 */ |
| 1499 core.List<core.String> archiveUris; | 1602 core.List<core.String> archiveUris; |
| 1500 /** | 1603 /** |
| 1501 * [Optional] The arguments to pass to the driver. Do not include arguments, | 1604 * Optional The arguments to pass to the driver. Do not include arguments, |
| 1502 * such as `-libjars` or `-Dfoo=bar`, that can be set as job properties, since | 1605 * such as -libjars or -Dfoo=bar, that can be set as job properties, since a |
| 1503 * a collision may occur that causes an incorrect job submission. | 1606 * collision may occur that causes an incorrect job submission. |
| 1504 */ | 1607 */ |
| 1505 core.List<core.String> args; | 1608 core.List<core.String> args; |
| 1506 /** | 1609 /** |
| 1507 * [Optional] HCFS URIs of files to be copied to the working directory of | 1610 * Optional HCFS URIs of files to be copied to the working directory of Hadoop |
| 1508 * Hadoop drivers and distributed tasks. Useful for naively parallel tasks. | 1611 * drivers and distributed tasks. Useful for naively parallel tasks. |
| 1509 */ | 1612 */ |
| 1510 core.List<core.String> fileUris; | 1613 core.List<core.String> fileUris; |
| 1511 /** | 1614 /** |
| 1512 * [Optional] Jar file URIs to add to the CLASSPATHs of the Hadoop driver and | 1615 * Optional Jar file URIs to add to the CLASSPATHs of the Hadoop driver and |
| 1513 * tasks. | 1616 * tasks. |
| 1514 */ | 1617 */ |
| 1515 core.List<core.String> jarFileUris; | 1618 core.List<core.String> jarFileUris; |
| 1516 /** [Optional] The runtime log configuration for job execution. */ | 1619 /** Optional The runtime log configuration for job execution. */ |
| 1517 LoggingConfiguration loggingConfiguration; | 1620 LoggingConfiguration loggingConfiguration; |
| 1518 /** | 1621 /** |
| 1519 * The name of the driver's main class. The jar file containing the class must | 1622 * The name of the driver's main class. The jar file containing the class must |
| 1520 * be in the default CLASSPATH or specified in `jar_file_uris`. | 1623 * be in the default CLASSPATH or specified in jar_file_uris. |
| 1521 */ | 1624 */ |
| 1522 core.String mainClass; | 1625 core.String mainClass; |
| 1523 /** | 1626 /** |
| 1524 * The Hadoop Compatible Filesystem (HCFS) URI of the jar file containing the | 1627 * The Hadoop Compatible Filesystem (HCFS) URI of the jar file containing the |
| 1525 * main class. Examples: | 1628 * main class. Examples: |
| 1526 * 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' | 1629 * 'gs://foo-bucket/analytics-binaries/extract-useful-metrics-mr.jar' |
| 1527 * 'hdfs:/tmp/test-samples/custom-wordcount.jar' | 1630 * 'hdfs:/tmp/test-samples/custom-wordcount.jar' |
| 1528 * 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' | 1631 * 'file:///home/usr/lib/hadoop-mapreduce/hadoop-mapreduce-examples.jar' |
| 1529 */ | 1632 */ |
| 1530 core.String mainJarFileUri; | 1633 core.String mainJarFileUri; |
| 1531 /** | 1634 /** |
| 1532 * [Optional] A mapping of property names to values, used to configure Hadoop. | 1635 * Optional A mapping of property names to values, used to configure Hadoop. |
| 1533 * Properties that conflict with values set by the Cloud Dataproc API may be | 1636 * Properties that conflict with values set by the Cloud Dataproc API may be |
| 1534 * overwritten. Can include properties set in /etc/hadoop/conf / * -site and | 1637 * overwritten. Can include properties set in /etc/hadoop/conf / * -site and |
| 1535 * classes in user code. | 1638 * classes in user code. |
| 1536 */ | 1639 */ |
| 1537 core.Map<core.String, core.String> properties; | 1640 core.Map<core.String, core.String> properties; |
| 1538 | 1641 |
| 1539 HadoopJob(); | 1642 HadoopJob(); |
| 1540 | 1643 |
| 1541 HadoopJob.fromJson(core.Map _json) { | 1644 HadoopJob.fromJson(core.Map _json) { |
| 1542 if (_json.containsKey("archiveUris")) { | 1645 if (_json.containsKey("archiveUris")) { |
| (...skipping 48 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1591 if (properties != null) { | 1694 if (properties != null) { |
| 1592 _json["properties"] = properties; | 1695 _json["properties"] = properties; |
| 1593 } | 1696 } |
| 1594 return _json; | 1697 return _json; |
| 1595 } | 1698 } |
| 1596 } | 1699 } |
| 1597 | 1700 |
| 1598 /** A Cloud Dataproc job for running Hive queries on YARN. */ | 1701 /** A Cloud Dataproc job for running Hive queries on YARN. */ |
| 1599 class HiveJob { | 1702 class HiveJob { |
| 1600 /** | 1703 /** |
| 1601 * [Optional] Whether to continue executing queries if a query fails. The | 1704 * Optional Whether to continue executing queries if a query fails. The |
| 1602 * default value is `false`. Setting to `true` can be useful when executing | 1705 * default value is false. Setting to true can be useful when executing |
| 1603 * independent parallel queries. | 1706 * independent parallel queries. |
| 1604 */ | 1707 */ |
| 1605 core.bool continueOnFailure; | 1708 core.bool continueOnFailure; |
| 1606 /** | 1709 /** |
| 1607 * [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Hive | 1710 * Optional HCFS URIs of jar files to add to the CLASSPATH of the Hive server |
| 1608 * server and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs. | 1711 * and Hadoop MapReduce (MR) tasks. Can contain Hive SerDes and UDFs. |
| 1609 */ | 1712 */ |
| 1610 core.List<core.String> jarFileUris; | 1713 core.List<core.String> jarFileUris; |
| 1611 /** | 1714 /** |
| 1612 * [Optional] A mapping of property names and values, used to configure Hive. | 1715 * Optional A mapping of property names and values, used to configure Hive. |
| 1613 * Properties that conflict with values set by the Cloud Dataproc API may be | 1716 * Properties that conflict with values set by the Cloud Dataproc API may be |
| 1614 * overwritten. Can include properties set in /etc/hadoop/conf / * -site.xml, | 1717 * overwritten. Can include properties set in /etc/hadoop/conf / * -site.xml, |
| 1615 * /etc/hive/conf/hive-site.xml, and classes in user code. | 1718 * /etc/hive/conf/hive-site.xml, and classes in user code. |
| 1616 */ | 1719 */ |
| 1617 core.Map<core.String, core.String> properties; | 1720 core.Map<core.String, core.String> properties; |
| 1618 /** The HCFS URI of the script that contains Hive queries. */ | 1721 /** The HCFS URI of the script that contains Hive queries. */ |
| 1619 core.String queryFileUri; | 1722 core.String queryFileUri; |
| 1620 /** A list of queries. */ | 1723 /** A list of queries. */ |
| 1621 QueryList queryList; | 1724 QueryList queryList; |
| 1622 /** | 1725 /** |
| 1623 * [Optional] Mapping of query variable names to values (equivalent to the | 1726 * Optional Mapping of query variable names to values (equivalent to the Hive |
| 1624 * Hive command: `SET name="value";`). | 1727 * command: SET name="value";). |
| 1625 */ | 1728 */ |
| 1626 core.Map<core.String, core.String> scriptVariables; | 1729 core.Map<core.String, core.String> scriptVariables; |
| 1627 | 1730 |
| 1628 HiveJob(); | 1731 HiveJob(); |
| 1629 | 1732 |
| 1630 HiveJob.fromJson(core.Map _json) { | 1733 HiveJob.fromJson(core.Map _json) { |
| 1631 if (_json.containsKey("continueOnFailure")) { | 1734 if (_json.containsKey("continueOnFailure")) { |
| 1632 continueOnFailure = _json["continueOnFailure"]; | 1735 continueOnFailure = _json["continueOnFailure"]; |
| 1633 } | 1736 } |
| 1634 if (_json.containsKey("jarFileUris")) { | 1737 if (_json.containsKey("jarFileUris")) { |
| (...skipping 38 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1673 } | 1776 } |
| 1674 | 1777 |
| 1675 /** | 1778 /** |
| 1676 * The configuration settings for Google Compute Engine resources in an instance | 1779 * The configuration settings for Google Compute Engine resources in an instance |
| 1677 * group, such as a master or worker group. | 1780 * group, such as a master or worker group. |
| 1678 */ | 1781 */ |
| 1679 class InstanceGroupConfiguration { | 1782 class InstanceGroupConfiguration { |
| 1680 /** Disk option configuration settings. */ | 1783 /** Disk option configuration settings. */ |
| 1681 DiskConfiguration diskConfiguration; | 1784 DiskConfiguration diskConfiguration; |
| 1682 /** | 1785 /** |
| 1683 * [Output-only] The Google Compute Engine image resource used for cluster | 1786 * Output-only The Google Compute Engine image resource used for cluster |
| 1684 * instances. Inferred from `SoftwareConfiguration.image_version`. | 1787 * instances. Inferred from SoftwareConfiguration.image_version. |
| 1685 */ | 1788 */ |
| 1686 core.String imageUri; | 1789 core.String imageUri; |
| 1687 /** | 1790 /** |
| 1688 * The list of instance names. Dataproc derives the names from `cluster_name`, | 1791 * The list of instance names. Dataproc derives the names from cluster_name, |
| 1689 * `num_instances`, and the instance group if not set by user (recommended | 1792 * num_instances, and the instance group if not set by user (recommended |
| 1690 * practice is to let Dataproc derive the name). | 1793 * practice is to let Dataproc derive the name). |
| 1691 */ | 1794 */ |
| 1692 core.List<core.String> instanceNames; | 1795 core.List<core.String> instanceNames; |
| 1693 /** Specifies that this instance group contains Preemptible Instances. */ | 1796 /** Specifies that this instance group contains Preemptible Instances. */ |
| 1694 core.bool isPreemptible; | 1797 core.bool isPreemptible; |
| 1695 /** | 1798 /** |
| 1696 * The Google Compute Engine machine type used for cluster instances. Example: | 1799 * The Google Compute Engine machine type used for cluster instances. Example: |
| 1697 * `https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1
-a/machineTypes/n1-standard-2`. | 1800 * https://www.googleapis.com/compute/v1/projects/[project_id]/zones/us-east1-
a/machineTypes/n1-standard-2. |
| 1698 */ | 1801 */ |
| 1699 core.String machineTypeUri; | 1802 core.String machineTypeUri; |
| 1700 /** | 1803 /** |
| 1701 * [Output-only] The configuration for Google Compute Engine Instance Group | 1804 * Output-only The configuration for Google Compute Engine Instance Group |
| 1702 * Manager that manages this group. This is only used for preemptible instance | 1805 * Manager that manages this group. This is only used for preemptible instance |
| 1703 * groups. | 1806 * groups. |
| 1704 */ | 1807 */ |
| 1705 ManagedGroupConfiguration managedGroupConfiguration; | 1808 ManagedGroupConfiguration managedGroupConfiguration; |
| 1706 /** | 1809 /** |
| 1707 * The number of VM instances in the instance group. For master instance | 1810 * The number of VM instances in the instance group. For master instance |
| 1708 * groups, must be set to 1. | 1811 * groups, must be set to 1. |
| 1709 */ | 1812 */ |
| 1710 core.int numInstances; | 1813 core.int numInstances; |
| 1711 | 1814 |
| (...skipping 46 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1758 if (numInstances != null) { | 1861 if (numInstances != null) { |
| 1759 _json["numInstances"] = numInstances; | 1862 _json["numInstances"] = numInstances; |
| 1760 } | 1863 } |
| 1761 return _json; | 1864 return _json; |
| 1762 } | 1865 } |
| 1763 } | 1866 } |
| 1764 | 1867 |
| 1765 /** A Cloud Dataproc job resource. */ | 1868 /** A Cloud Dataproc job resource. */ |
| 1766 class Job { | 1869 class Job { |
| 1767 /** | 1870 /** |
| 1768 * [Output-only] If present, the location of miscellaneous control files which | 1871 * Output-only If present, the location of miscellaneous control files which |
| 1769 * may be used as part of job setup and handling. If not present, control | 1872 * may be used as part of job setup and handling. If not present, control |
| 1770 * files may be placed in the same location as `driver_output_uri`. | 1873 * files may be placed in the same location as driver_output_uri. |
| 1771 */ | 1874 */ |
| 1772 core.String driverControlFilesUri; | 1875 core.String driverControlFilesUri; |
| 1773 /** | 1876 /** |
| 1774 * [Output-only] A URI pointing to the location of the stdin of the job's | 1877 * Output-only A URI pointing to the location of the stdin of the job's driver |
| 1775 * driver program, only set if the job is interactive. | 1878 * program, only set if the job is interactive. |
| 1776 */ | 1879 */ |
| 1777 core.String driverInputResourceUri; | 1880 core.String driverInputResourceUri; |
| 1778 /** | 1881 /** |
| 1779 * [Output-only] A URI pointing to the location of the stdout of the job's | 1882 * Output-only A URI pointing to the location of the stdout of the job's |
| 1780 * driver program. | 1883 * driver program. |
| 1781 */ | 1884 */ |
| 1782 core.String driverOutputResourceUri; | 1885 core.String driverOutputResourceUri; |
| 1783 /** Job is a Hadoop job. */ | 1886 /** Job is a Hadoop job. */ |
| 1784 HadoopJob hadoopJob; | 1887 HadoopJob hadoopJob; |
| 1785 /** Job is a Hive job. */ | 1888 /** Job is a Hive job. */ |
| 1786 HiveJob hiveJob; | 1889 HiveJob hiveJob; |
| 1787 /** | 1890 /** |
| 1788 * [Optional] If set to `true`, the driver's stdin will be kept open and | 1891 * Optional If set to true, the driver's stdin will be kept open and |
| 1789 * `driver_input_uri` will be set to provide a path at which additional input | 1892 * driver_input_uri will be set to provide a path at which additional input |
| 1790 * can be sent to the driver. | 1893 * can be sent to the driver. |
| 1791 */ | 1894 */ |
| 1792 core.bool interactive; | 1895 core.bool interactive; |
| 1793 /** | 1896 /** |
| 1794 * [Optional] The labels to associate with this job. Label keys must be | 1897 * Optional The labels to associate with this job.Label keys must be between 1 |
| 1795 * between 1 and 63 characters long, and must conform to the following regular | 1898 * and 63 characters long, and must conform to the following regular |
| 1796 * expression: \p{Ll}\p{Lo}{0,62} Label values must be between 1 and 63 | 1899 * expression: \p{Ll}\p{Lo}{0,62}Label values must be between 1 and 63 |
| 1797 * characters long, and must conform to the following regular expression: | 1900 * characters long, and must conform to the following regular expression: |
| 1798 * [\p{Ll}\p{Lo}\p{N}_-]{0,63} No more than 64 labels can be associated with a | 1901 * \p{Ll}\p{Lo}\p{N}_-{0,63}No more than 64 labels can be associated with a |
| 1799 * given job. | 1902 * given job. |
| 1800 */ | 1903 */ |
| 1801 core.Map<core.String, core.String> labels; | 1904 core.Map<core.String, core.String> labels; |
| 1802 /** Job is a Pig job. */ | 1905 /** Job is a Pig job. */ |
| 1803 PigJob pigJob; | 1906 PigJob pigJob; |
| 1804 /** | 1907 /** |
| 1805 * [Required] Job information, including how, when, and where to run the job. | 1908 * Required Job information, including how, when, and where to run the job. |
| 1806 */ | 1909 */ |
| 1807 JobPlacement placement; | 1910 JobPlacement placement; |
| 1808 /** Job is a Pyspark job. */ | 1911 /** Job is a Pyspark job. */ |
| 1809 PySparkJob pysparkJob; | 1912 PySparkJob pysparkJob; |
| 1810 /** | 1913 /** |
| 1811 * [Optional] The fully qualified reference to the job, which can be used to | 1914 * Optional The fully qualified reference to the job, which can be used to |
| 1812 * obtain the equivalent REST path of the job resource. If this property is | 1915 * obtain the equivalent REST path of the job resource. If this property is |
| 1813 * not specified when a job is created, the server generates a job_id. | 1916 * not specified when a job is created, the server generates a |
| 1917 * <code>job_id</code>. |
| 1814 */ | 1918 */ |
| 1815 JobReference reference; | 1919 JobReference reference; |
| 1920 /** Optional Job scheduling configuration. */ |
| 1921 JobScheduling scheduling; |
| 1816 /** Job is a Spark job. */ | 1922 /** Job is a Spark job. */ |
| 1817 SparkJob sparkJob; | 1923 SparkJob sparkJob; |
| 1818 /** Job is a SparkSql job. */ | 1924 /** Job is a SparkSql job. */ |
| 1819 SparkSqlJob sparkSqlJob; | 1925 SparkSqlJob sparkSqlJob; |
| 1820 /** | 1926 /** |
| 1821 * [Output-only] The job status. Additional application-specific status | 1927 * Output-only The job status. Additional application-specific status |
| 1822 * information may be contained in the type_job and yarn_applications fields. | 1928 * information may be contained in the <code>type_job</code> and |
| 1929 * <code>yarn_applications</code> fields. |
| 1823 */ | 1930 */ |
| 1824 JobStatus status; | 1931 JobStatus status; |
| 1825 /** [Output-only] The previous job status. */ | 1932 /** Output-only The previous job status. */ |
| 1826 core.List<JobStatus> statusHistory; | 1933 core.List<JobStatus> statusHistory; |
| 1827 /** | 1934 /** |
| 1828 * [Output-only] The email address of the user submitting the job. For jobs | 1935 * Output-only The email address of the user submitting the job. For jobs |
| 1829 * submitted on the cluster, the address is username@hostname. | 1936 * submitted on the cluster, the address is <code>username@hostname</code>. |
| 1830 */ | 1937 */ |
| 1831 core.String submittedBy; | 1938 core.String submittedBy; |
| 1832 /** [Output-only] The collection of YARN applications spun up by this job. */ | 1939 /** Output-only The collection of YARN applications spun up by this job. */ |
| 1833 core.List<YarnApplication> yarnApplications; | 1940 core.List<YarnApplication> yarnApplications; |
| 1834 | 1941 |
| 1835 Job(); | 1942 Job(); |
| 1836 | 1943 |
| 1837 Job.fromJson(core.Map _json) { | 1944 Job.fromJson(core.Map _json) { |
| 1838 if (_json.containsKey("driverControlFilesUri")) { | 1945 if (_json.containsKey("driverControlFilesUri")) { |
| 1839 driverControlFilesUri = _json["driverControlFilesUri"]; | 1946 driverControlFilesUri = _json["driverControlFilesUri"]; |
| 1840 } | 1947 } |
| 1841 if (_json.containsKey("driverInputResourceUri")) { | 1948 if (_json.containsKey("driverInputResourceUri")) { |
| 1842 driverInputResourceUri = _json["driverInputResourceUri"]; | 1949 driverInputResourceUri = _json["driverInputResourceUri"]; |
| (...skipping 18 matching lines...) Expand all Loading... |
| 1861 } | 1968 } |
| 1862 if (_json.containsKey("placement")) { | 1969 if (_json.containsKey("placement")) { |
| 1863 placement = new JobPlacement.fromJson(_json["placement"]); | 1970 placement = new JobPlacement.fromJson(_json["placement"]); |
| 1864 } | 1971 } |
| 1865 if (_json.containsKey("pysparkJob")) { | 1972 if (_json.containsKey("pysparkJob")) { |
| 1866 pysparkJob = new PySparkJob.fromJson(_json["pysparkJob"]); | 1973 pysparkJob = new PySparkJob.fromJson(_json["pysparkJob"]); |
| 1867 } | 1974 } |
| 1868 if (_json.containsKey("reference")) { | 1975 if (_json.containsKey("reference")) { |
| 1869 reference = new JobReference.fromJson(_json["reference"]); | 1976 reference = new JobReference.fromJson(_json["reference"]); |
| 1870 } | 1977 } |
| 1978 if (_json.containsKey("scheduling")) { |
| 1979 scheduling = new JobScheduling.fromJson(_json["scheduling"]); |
| 1980 } |
| 1871 if (_json.containsKey("sparkJob")) { | 1981 if (_json.containsKey("sparkJob")) { |
| 1872 sparkJob = new SparkJob.fromJson(_json["sparkJob"]); | 1982 sparkJob = new SparkJob.fromJson(_json["sparkJob"]); |
| 1873 } | 1983 } |
| 1874 if (_json.containsKey("sparkSqlJob")) { | 1984 if (_json.containsKey("sparkSqlJob")) { |
| 1875 sparkSqlJob = new SparkSqlJob.fromJson(_json["sparkSqlJob"]); | 1985 sparkSqlJob = new SparkSqlJob.fromJson(_json["sparkSqlJob"]); |
| 1876 } | 1986 } |
| 1877 if (_json.containsKey("status")) { | 1987 if (_json.containsKey("status")) { |
| 1878 status = new JobStatus.fromJson(_json["status"]); | 1988 status = new JobStatus.fromJson(_json["status"]); |
| 1879 } | 1989 } |
| 1880 if (_json.containsKey("statusHistory")) { | 1990 if (_json.containsKey("statusHistory")) { |
| (...skipping 35 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 1916 } | 2026 } |
| 1917 if (placement != null) { | 2027 if (placement != null) { |
| 1918 _json["placement"] = (placement).toJson(); | 2028 _json["placement"] = (placement).toJson(); |
| 1919 } | 2029 } |
| 1920 if (pysparkJob != null) { | 2030 if (pysparkJob != null) { |
| 1921 _json["pysparkJob"] = (pysparkJob).toJson(); | 2031 _json["pysparkJob"] = (pysparkJob).toJson(); |
| 1922 } | 2032 } |
| 1923 if (reference != null) { | 2033 if (reference != null) { |
| 1924 _json["reference"] = (reference).toJson(); | 2034 _json["reference"] = (reference).toJson(); |
| 1925 } | 2035 } |
| 2036 if (scheduling != null) { |
| 2037 _json["scheduling"] = (scheduling).toJson(); |
| 2038 } |
| 1926 if (sparkJob != null) { | 2039 if (sparkJob != null) { |
| 1927 _json["sparkJob"] = (sparkJob).toJson(); | 2040 _json["sparkJob"] = (sparkJob).toJson(); |
| 1928 } | 2041 } |
| 1929 if (sparkSqlJob != null) { | 2042 if (sparkSqlJob != null) { |
| 1930 _json["sparkSqlJob"] = (sparkSqlJob).toJson(); | 2043 _json["sparkSqlJob"] = (sparkSqlJob).toJson(); |
| 1931 } | 2044 } |
| 1932 if (status != null) { | 2045 if (status != null) { |
| 1933 _json["status"] = (status).toJson(); | 2046 _json["status"] = (status).toJson(); |
| 1934 } | 2047 } |
| 1935 if (statusHistory != null) { | 2048 if (statusHistory != null) { |
| 1936 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); | 2049 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); |
| 1937 } | 2050 } |
| 1938 if (submittedBy != null) { | 2051 if (submittedBy != null) { |
| 1939 _json["submittedBy"] = submittedBy; | 2052 _json["submittedBy"] = submittedBy; |
| 1940 } | 2053 } |
| 1941 if (yarnApplications != null) { | 2054 if (yarnApplications != null) { |
| 1942 _json["yarnApplications"] = yarnApplications.map((value) => (value).toJson
()).toList(); | 2055 _json["yarnApplications"] = yarnApplications.map((value) => (value).toJson
()).toList(); |
| 1943 } | 2056 } |
| 1944 return _json; | 2057 return _json; |
| 1945 } | 2058 } |
| 1946 } | 2059 } |
| 1947 | 2060 |
| 1948 /** Cloud Dataproc job configuration. */ | 2061 /** Cloud Dataproc job configuration. */ |
| 1949 class JobPlacement { | 2062 class JobPlacement { |
| 1950 /** [Required] The name of the cluster where the job will be submitted. */ | 2063 /** Required The name of the cluster where the job will be submitted. */ |
| 1951 core.String clusterName; | 2064 core.String clusterName; |
| 1952 /** | 2065 /** |
| 1953 * [Output-only] A cluster UUID generated by the Dataproc service when the job | 2066 * Output-only A cluster UUID generated by the Dataproc service when the job |
| 1954 * is submitted. | 2067 * is submitted. |
| 1955 */ | 2068 */ |
| 1956 core.String clusterUuid; | 2069 core.String clusterUuid; |
| 1957 | 2070 |
| 1958 JobPlacement(); | 2071 JobPlacement(); |
| 1959 | 2072 |
| 1960 JobPlacement.fromJson(core.Map _json) { | 2073 JobPlacement.fromJson(core.Map _json) { |
| 1961 if (_json.containsKey("clusterName")) { | 2074 if (_json.containsKey("clusterName")) { |
| 1962 clusterName = _json["clusterName"]; | 2075 clusterName = _json["clusterName"]; |
| 1963 } | 2076 } |
| (...skipping 10 matching lines...) Expand all Loading... |
| 1974 if (clusterUuid != null) { | 2087 if (clusterUuid != null) { |
| 1975 _json["clusterUuid"] = clusterUuid; | 2088 _json["clusterUuid"] = clusterUuid; |
| 1976 } | 2089 } |
| 1977 return _json; | 2090 return _json; |
| 1978 } | 2091 } |
| 1979 } | 2092 } |
| 1980 | 2093 |
| 1981 /** Encapsulates the full scoping used to reference a job. */ | 2094 /** Encapsulates the full scoping used to reference a job. */ |
| 1982 class JobReference { | 2095 class JobReference { |
| 1983 /** | 2096 /** |
| 1984 * [Required] The job ID, which must be unique within the project. The job ID | 2097 * Required The job ID, which must be unique within the project. The job ID is |
| 1985 * is generated by the server upon job submission or provided by the user as a | 2098 * generated by the server upon job submission or provided by the user as a |
| 1986 * means to perform retries without creating duplicate jobs. The ID must | 2099 * means to perform retries without creating duplicate jobs. The ID must |
| 1987 * contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens | 2100 * contain only letters (a-z, A-Z), numbers (0-9), underscores (_), or hyphens |
| 1988 * (-). The maximum length is 512 characters. | 2101 * (-). The maximum length is 512 characters. |
| 1989 */ | 2102 */ |
| 1990 core.String jobId; | 2103 core.String jobId; |
| 1991 /** | 2104 /** |
| 1992 * [Required] The ID of the Google Cloud Platform project that the job belongs | 2105 * Required The ID of the Google Cloud Platform project that the job belongs |
| 1993 * to. | 2106 * to. |
| 1994 */ | 2107 */ |
| 1995 core.String projectId; | 2108 core.String projectId; |
| 1996 | 2109 |
| 1997 JobReference(); | 2110 JobReference(); |
| 1998 | 2111 |
| 1999 JobReference.fromJson(core.Map _json) { | 2112 JobReference.fromJson(core.Map _json) { |
| 2000 if (_json.containsKey("jobId")) { | 2113 if (_json.containsKey("jobId")) { |
| 2001 jobId = _json["jobId"]; | 2114 jobId = _json["jobId"]; |
| 2002 } | 2115 } |
| 2003 if (_json.containsKey("projectId")) { | 2116 if (_json.containsKey("projectId")) { |
| 2004 projectId = _json["projectId"]; | 2117 projectId = _json["projectId"]; |
| 2005 } | 2118 } |
| 2006 } | 2119 } |
| 2007 | 2120 |
| 2008 core.Map toJson() { | 2121 core.Map toJson() { |
| 2009 var _json = new core.Map(); | 2122 var _json = new core.Map(); |
| 2010 if (jobId != null) { | 2123 if (jobId != null) { |
| 2011 _json["jobId"] = jobId; | 2124 _json["jobId"] = jobId; |
| 2012 } | 2125 } |
| 2013 if (projectId != null) { | 2126 if (projectId != null) { |
| 2014 _json["projectId"] = projectId; | 2127 _json["projectId"] = projectId; |
| 2015 } | 2128 } |
| 2016 return _json; | 2129 return _json; |
| 2017 } | 2130 } |
| 2018 } | 2131 } |
| 2019 | 2132 |
| 2133 /** |
| 2134 * Job scheduling options.Beta Feature: These options are available for testing |
| 2135 * purposes only. They may be changed before final release. |
| 2136 */ |
| 2137 class JobScheduling { |
| 2138 /** |
| 2139 * Optional Maximum number of times per hour a driver may be restarted as a |
| 2140 * result of driver terminating with non-zero code before job is reported |
| 2141 * failed.A job may be reported as thrashing if driver exits with non-zero |
| 2142 * code 4 times within 10 minute window.Maximum value is 10. |
| 2143 */ |
| 2144 core.int maxFailuresPerHour; |
| 2145 |
| 2146 JobScheduling(); |
| 2147 |
| 2148 JobScheduling.fromJson(core.Map _json) { |
| 2149 if (_json.containsKey("maxFailuresPerHour")) { |
| 2150 maxFailuresPerHour = _json["maxFailuresPerHour"]; |
| 2151 } |
| 2152 } |
| 2153 |
| 2154 core.Map toJson() { |
| 2155 var _json = new core.Map(); |
| 2156 if (maxFailuresPerHour != null) { |
| 2157 _json["maxFailuresPerHour"] = maxFailuresPerHour; |
| 2158 } |
| 2159 return _json; |
| 2160 } |
| 2161 } |
| 2162 |
| 2020 /** Cloud Dataproc job status. */ | 2163 /** Cloud Dataproc job status. */ |
| 2021 class JobStatus { | 2164 class JobStatus { |
| 2022 /** | 2165 /** |
| 2023 * [Optional] Job state details, such as an error description if the state is | 2166 * Optional Job state details, such as an error description if the state is |
| 2024 * ERROR. | 2167 * <code>ERROR</code>. |
| 2025 */ | 2168 */ |
| 2026 core.String details; | 2169 core.String details; |
| 2027 /** | 2170 /** |
| 2028 * [Required] A state message specifying the overall job state. | 2171 * Required A state message specifying the overall job state. |
| 2029 * Possible string values are: | 2172 * Possible string values are: |
| 2030 * - "STATE_UNSPECIFIED" : A STATE_UNSPECIFIED. | 2173 * - "STATE_UNSPECIFIED" : The job state is unknown. |
| 2031 * - "PENDING" : A PENDING. | 2174 * - "PENDING" : The job is pending; it has been submitted, but is not yet |
| 2032 * - "SETUP_DONE" : A SETUP_DONE. | 2175 * running. |
| 2033 * - "RUNNING" : A RUNNING. | 2176 * - "SETUP_DONE" : Job has been received by the service and completed initial |
| 2034 * - "CANCEL_PENDING" : A CANCEL_PENDING. | 2177 * setup; it will shortly be submitted to the cluster. |
| 2035 * - "CANCEL_STARTED" : A CANCEL_STARTED. | 2178 * - "RUNNING" : The job is running on the cluster. |
| 2036 * - "CANCELLED" : A CANCELLED. | 2179 * - "CANCEL_PENDING" : A CancelJob request has been received, but is pending. |
| 2037 * - "DONE" : A DONE. | 2180 * - "CANCEL_STARTED" : Transient in-flight resources have been canceled, and |
| 2038 * - "ERROR" : A ERROR. | 2181 * the request to cancel the running job has been issued to the cluster. |
| 2182 * - "CANCELLED" : The job cancelation was successful. |
| 2183 * - "DONE" : The job has completed successfully. |
| 2184 * - "ERROR" : The job has completed, but encountered an error. |
| 2185 * - "ATTEMPT_FAILURE" : Job attempt has failed. The detail field contains |
| 2186 * failure details for this attempt.Applies to restartable jobs only. |
| 2039 */ | 2187 */ |
| 2040 core.String state; | 2188 core.String state; |
| 2041 /** [Output-only] The time when this state was entered. */ | 2189 /** Output-only The time when this state was entered. */ |
| 2042 core.String stateStartTime; | 2190 core.String stateStartTime; |
| 2043 | 2191 |
| 2044 JobStatus(); | 2192 JobStatus(); |
| 2045 | 2193 |
| 2046 JobStatus.fromJson(core.Map _json) { | 2194 JobStatus.fromJson(core.Map _json) { |
| 2047 if (_json.containsKey("details")) { | 2195 if (_json.containsKey("details")) { |
| 2048 details = _json["details"]; | 2196 details = _json["details"]; |
| 2049 } | 2197 } |
| 2050 if (_json.containsKey("state")) { | 2198 if (_json.containsKey("state")) { |
| 2051 state = _json["state"]; | 2199 state = _json["state"]; |
| (...skipping 13 matching lines...) Expand all Loading... |
| 2065 } | 2213 } |
| 2066 if (stateStartTime != null) { | 2214 if (stateStartTime != null) { |
| 2067 _json["stateStartTime"] = stateStartTime; | 2215 _json["stateStartTime"] = stateStartTime; |
| 2068 } | 2216 } |
| 2069 return _json; | 2217 return _json; |
| 2070 } | 2218 } |
| 2071 } | 2219 } |
| 2072 | 2220 |
| 2073 /** The list of all clusters in a project. */ | 2221 /** The list of all clusters in a project. */ |
| 2074 class ListClustersResponse { | 2222 class ListClustersResponse { |
| 2075 /** [Output-only] The clusters in the project. */ | 2223 /** Output-only The clusters in the project. */ |
| 2076 core.List<Cluster> clusters; | 2224 core.List<Cluster> clusters; |
| 2077 /** The standard List next-page token. */ | 2225 /** The standard List next-page token. */ |
| 2078 core.String nextPageToken; | 2226 core.String nextPageToken; |
| 2079 | 2227 |
| 2080 ListClustersResponse(); | 2228 ListClustersResponse(); |
| 2081 | 2229 |
| 2082 ListClustersResponse.fromJson(core.Map _json) { | 2230 ListClustersResponse.fromJson(core.Map _json) { |
| 2083 if (_json.containsKey("clusters")) { | 2231 if (_json.containsKey("clusters")) { |
| 2084 clusters = _json["clusters"].map((value) => new Cluster.fromJson(value)).t
oList(); | 2232 clusters = _json["clusters"].map((value) => new Cluster.fromJson(value)).t
oList(); |
| 2085 } | 2233 } |
| 2086 if (_json.containsKey("nextPageToken")) { | 2234 if (_json.containsKey("nextPageToken")) { |
| 2087 nextPageToken = _json["nextPageToken"]; | 2235 nextPageToken = _json["nextPageToken"]; |
| 2088 } | 2236 } |
| 2089 } | 2237 } |
| 2090 | 2238 |
| 2091 core.Map toJson() { | 2239 core.Map toJson() { |
| 2092 var _json = new core.Map(); | 2240 var _json = new core.Map(); |
| 2093 if (clusters != null) { | 2241 if (clusters != null) { |
| 2094 _json["clusters"] = clusters.map((value) => (value).toJson()).toList(); | 2242 _json["clusters"] = clusters.map((value) => (value).toJson()).toList(); |
| 2095 } | 2243 } |
| 2096 if (nextPageToken != null) { | 2244 if (nextPageToken != null) { |
| 2097 _json["nextPageToken"] = nextPageToken; | 2245 _json["nextPageToken"] = nextPageToken; |
| 2098 } | 2246 } |
| 2099 return _json; | 2247 return _json; |
| 2100 } | 2248 } |
| 2101 } | 2249 } |
| 2102 | 2250 |
| 2103 /** A list of jobs in a project. */ | 2251 /** A list of jobs in a project. */ |
| 2104 class ListJobsResponse { | 2252 class ListJobsResponse { |
| 2105 /** [Output-only] Jobs list. */ | 2253 /** Output-only Jobs list. */ |
| 2106 core.List<Job> jobs; | 2254 core.List<Job> jobs; |
| 2107 /** | 2255 /** |
| 2108 * [Optional] This token is included in the response if there are more results | 2256 * Optional This token is included in the response if there are more results |
| 2109 * to fetch. To fetch additional results, provide this value as the | 2257 * to fetch. To fetch additional results, provide this value as the page_token |
| 2110 * `page_token` in a subsequent ListJobsRequest. | 2258 * in a subsequent <code>ListJobsRequest</code>. |
| 2111 */ | 2259 */ |
| 2112 core.String nextPageToken; | 2260 core.String nextPageToken; |
| 2113 | 2261 |
| 2114 ListJobsResponse(); | 2262 ListJobsResponse(); |
| 2115 | 2263 |
| 2116 ListJobsResponse.fromJson(core.Map _json) { | 2264 ListJobsResponse.fromJson(core.Map _json) { |
| 2117 if (_json.containsKey("jobs")) { | 2265 if (_json.containsKey("jobs")) { |
| 2118 jobs = _json["jobs"].map((value) => new Job.fromJson(value)).toList(); | 2266 jobs = _json["jobs"].map((value) => new Job.fromJson(value)).toList(); |
| 2119 } | 2267 } |
| 2120 if (_json.containsKey("nextPageToken")) { | 2268 if (_json.containsKey("nextPageToken")) { |
| (...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2161 _json["operations"] = operations.map((value) => (value).toJson()).toList()
; | 2309 _json["operations"] = operations.map((value) => (value).toJson()).toList()
; |
| 2162 } | 2310 } |
| 2163 return _json; | 2311 return _json; |
| 2164 } | 2312 } |
| 2165 } | 2313 } |
| 2166 | 2314 |
| 2167 /** The runtime logging configuration of the job. */ | 2315 /** The runtime logging configuration of the job. */ |
| 2168 class LoggingConfiguration { | 2316 class LoggingConfiguration { |
| 2169 /** | 2317 /** |
| 2170 * The per-package log levels for the driver. This may include "root" package | 2318 * The per-package log levels for the driver. This may include "root" package |
| 2171 * name to configure rootLogger. Examples: 'com.google = FATAL', 'root = | 2319 * name to configure rootLogger. Examples: 'com.google = FATAL', 'root = |
| 2172 * INFO', 'org.apache = DEBUG' | 2320 * INFO', 'org.apache = DEBUG' |
| 2173 */ | 2321 */ |
| 2174 core.Map<core.String, core.String> driverLogLevels; | 2322 core.Map<core.String, core.String> driverLogLevels; |
| 2175 | 2323 |
| 2176 LoggingConfiguration(); | 2324 LoggingConfiguration(); |
| 2177 | 2325 |
| 2178 LoggingConfiguration.fromJson(core.Map _json) { | 2326 LoggingConfiguration.fromJson(core.Map _json) { |
| 2179 if (_json.containsKey("driverLogLevels")) { | 2327 if (_json.containsKey("driverLogLevels")) { |
| 2180 driverLogLevels = _json["driverLogLevels"]; | 2328 driverLogLevels = _json["driverLogLevels"]; |
| 2181 } | 2329 } |
| 2182 } | 2330 } |
| 2183 | 2331 |
| 2184 core.Map toJson() { | 2332 core.Map toJson() { |
| 2185 var _json = new core.Map(); | 2333 var _json = new core.Map(); |
| 2186 if (driverLogLevels != null) { | 2334 if (driverLogLevels != null) { |
| 2187 _json["driverLogLevels"] = driverLogLevels; | 2335 _json["driverLogLevels"] = driverLogLevels; |
| 2188 } | 2336 } |
| 2189 return _json; | 2337 return _json; |
| 2190 } | 2338 } |
| 2191 } | 2339 } |
| 2192 | 2340 |
| 2193 /** Specifies the resources used to actively manage an instance group. */ | 2341 /** Specifies the resources used to actively manage an instance group. */ |
| 2194 class ManagedGroupConfiguration { | 2342 class ManagedGroupConfiguration { |
| 2195 /** [Output-only] The name of the Instance Group Manager for this group. */ | 2343 /** Output-only The name of the Instance Group Manager for this group. */ |
| 2196 core.String instanceGroupManagerName; | 2344 core.String instanceGroupManagerName; |
| 2197 /** | 2345 /** |
| 2198 * [Output-only] The name of the Instance Template used for the Managed | 2346 * Output-only The name of the Instance Template used for the Managed Instance |
| 2199 * Instance Group. | 2347 * Group. |
| 2200 */ | 2348 */ |
| 2201 core.String instanceTemplateName; | 2349 core.String instanceTemplateName; |
| 2202 | 2350 |
| 2203 ManagedGroupConfiguration(); | 2351 ManagedGroupConfiguration(); |
| 2204 | 2352 |
| 2205 ManagedGroupConfiguration.fromJson(core.Map _json) { | 2353 ManagedGroupConfiguration.fromJson(core.Map _json) { |
| 2206 if (_json.containsKey("instanceGroupManagerName")) { | 2354 if (_json.containsKey("instanceGroupManagerName")) { |
| 2207 instanceGroupManagerName = _json["instanceGroupManagerName"]; | 2355 instanceGroupManagerName = _json["instanceGroupManagerName"]; |
| 2208 } | 2356 } |
| 2209 if (_json.containsKey("instanceTemplateName")) { | 2357 if (_json.containsKey("instanceTemplateName")) { |
| (...skipping 11 matching lines...) Expand all Loading... |
| 2221 } | 2369 } |
| 2222 return _json; | 2370 return _json; |
| 2223 } | 2371 } |
| 2224 } | 2372 } |
| 2225 | 2373 |
| 2226 /** | 2374 /** |
| 2227 * Specifies an executable to run on a fully configured node and a timeout | 2375 * Specifies an executable to run on a fully configured node and a timeout |
| 2228 * period for executable completion. | 2376 * period for executable completion. |
| 2229 */ | 2377 */ |
| 2230 class NodeInitializationAction { | 2378 class NodeInitializationAction { |
| 2231 /** [Required] Google Cloud Storage URI of executable file. */ | 2379 /** Required Google Cloud Storage URI of executable file. */ |
| 2232 core.String executableFile; | 2380 core.String executableFile; |
| 2233 /** | 2381 /** |
| 2234 * [Optional] Amount of time executable has to complete. Default is 10 | 2382 * Optional Amount of time executable has to complete. Default is 10 minutes. |
| 2235 * minutes. Cluster creation fails with an explanatory error message (the name | 2383 * Cluster creation fails with an explanatory error message (the name of the |
| 2236 * of the executable that caused the error and the exceeded timeout period) if | 2384 * executable that caused the error and the exceeded timeout period) if the |
| 2237 * the executable is not completed at end of the timeout period. | 2385 * executable is not completed at end of the timeout period. |
| 2238 */ | 2386 */ |
| 2239 core.String executionTimeout; | 2387 core.String executionTimeout; |
| 2240 | 2388 |
| 2241 NodeInitializationAction(); | 2389 NodeInitializationAction(); |
| 2242 | 2390 |
| 2243 NodeInitializationAction.fromJson(core.Map _json) { | 2391 NodeInitializationAction.fromJson(core.Map _json) { |
| 2244 if (_json.containsKey("executableFile")) { | 2392 if (_json.containsKey("executableFile")) { |
| 2245 executableFile = _json["executableFile"]; | 2393 executableFile = _json["executableFile"]; |
| 2246 } | 2394 } |
| 2247 if (_json.containsKey("executionTimeout")) { | 2395 if (_json.containsKey("executionTimeout")) { |
| (...skipping 12 matching lines...) Expand all Loading... |
| 2260 return _json; | 2408 return _json; |
| 2261 } | 2409 } |
| 2262 } | 2410 } |
| 2263 | 2411 |
| 2264 /** | 2412 /** |
| 2265 * This resource represents a long-running operation that is the result of a | 2413 * This resource represents a long-running operation that is the result of a |
| 2266 * network API call. | 2414 * network API call. |
| 2267 */ | 2415 */ |
| 2268 class Operation { | 2416 class Operation { |
| 2269 /** | 2417 /** |
| 2270 * If the value is `false`, it means the operation is still in progress. If | 2418 * If the value is false, it means the operation is still in progress. If |
| 2271 * true, the operation is completed, and either `error` or `response` is | 2419 * true, the operation is completed, and either error or response is |
| 2272 * available. | 2420 * available. |
| 2273 */ | 2421 */ |
| 2274 core.bool done; | 2422 core.bool done; |
| 2275 /** The error result of the operation in case of failure or cancellation. */ | 2423 /** The error result of the operation in case of failure or cancellation. */ |
| 2276 Status error; | 2424 Status error; |
| 2277 /** | 2425 /** |
| 2278 * Service-specific metadata associated with the operation. It typically | 2426 * Service-specific metadata associated with the operation. It typically |
| 2279 * contains progress information and common metadata such as create time. Some | 2427 * contains progress information and common metadata such as create time. Some |
| 2280 * services might not provide such metadata. Any method that returns a | 2428 * services might not provide such metadata. Any method that returns a |
| 2281 * long-running operation should document the metadata type, if any. | 2429 * long-running operation should document the metadata type, if any. |
| 2282 * | 2430 * |
| 2283 * The values for Object must be JSON objects. It can consist of `num`, | 2431 * The values for Object must be JSON objects. It can consist of `num`, |
| 2284 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2432 * `String`, `bool` and `null` as well as `Map` and `List` values. |
| 2285 */ | 2433 */ |
| 2286 core.Map<core.String, core.Object> metadata; | 2434 core.Map<core.String, core.Object> metadata; |
| 2287 /** | 2435 /** |
| 2288 * The server-assigned name, which is only unique within the same service that | 2436 * The server-assigned name, which is only unique within the same service that |
| 2289 * originally returns it. If you use the default HTTP mapping, the `name` | 2437 * originally returns it. If you use the default HTTP mapping, the name should |
| 2290 * should have the format of `operations/some/unique/name`. | 2438 * have the format of operations/some/unique/name. |
| 2291 */ | 2439 */ |
| 2292 core.String name; | 2440 core.String name; |
| 2293 /** | 2441 /** |
| 2294 * The normal response of the operation in case of success. If the original | 2442 * The normal response of the operation in case of success. If the original |
| 2295 * method returns no data on success, such as `Delete`, the response is | 2443 * method returns no data on success, such as Delete, the response is |
| 2296 * `google.protobuf.Empty`. If the original method is standard | 2444 * google.protobuf.Empty. If the original method is standard |
| 2297 * `Get`/`Create`/`Update`, the response should be the resource. For other | 2445 * Get/Create/Update, the response should be the resource. For other methods, |
| 2298 * methods, the response should have the type `XxxResponse`, where `Xxx` is | 2446 * the response should have the type XxxResponse, where Xxx is the original |
| 2299 * the original method name. For example, if the original method name is | 2447 * method name. For example, if the original method name is TakeSnapshot(), |
| 2300 * `TakeSnapshot()`, the inferred response type is `TakeSnapshotResponse`. | 2448 * the inferred response type is TakeSnapshotResponse. |
| 2301 * | 2449 * |
| 2302 * The values for Object must be JSON objects. It can consist of `num`, | 2450 * The values for Object must be JSON objects. It can consist of `num`, |
| 2303 * `String`, `bool` and `null` as well as `Map` and `List` values. | 2451 * `String`, `bool` and `null` as well as `Map` and `List` values. |
| 2304 */ | 2452 */ |
| 2305 core.Map<core.String, core.Object> response; | 2453 core.Map<core.String, core.Object> response; |
| 2306 | 2454 |
| 2307 Operation(); | 2455 Operation(); |
| 2308 | 2456 |
| 2309 Operation.fromJson(core.Map _json) { | 2457 Operation.fromJson(core.Map _json) { |
| 2310 if (_json.containsKey("done")) { | 2458 if (_json.containsKey("done")) { |
| (...skipping 33 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2344 return _json; | 2492 return _json; |
| 2345 } | 2493 } |
| 2346 } | 2494 } |
| 2347 | 2495 |
| 2348 /** Metadata describing the operation. */ | 2496 /** Metadata describing the operation. */ |
| 2349 class OperationMetadata { | 2497 class OperationMetadata { |
| 2350 /** Name of the cluster for the operation. */ | 2498 /** Name of the cluster for the operation. */ |
| 2351 core.String clusterName; | 2499 core.String clusterName; |
| 2352 /** Cluster UUId for the operation. */ | 2500 /** Cluster UUId for the operation. */ |
| 2353 core.String clusterUuid; | 2501 core.String clusterUuid; |
| 2354 /** [Output-only] Short description of operation. */ | 2502 /** Output-only Short description of operation. */ |
| 2355 core.String description; | 2503 core.String description; |
| 2356 /** A message containing any operation metadata details. */ | 2504 /** A message containing any operation metadata details. */ |
| 2357 core.String details; | 2505 core.String details; |
| 2358 /** The time that the operation completed. */ | 2506 /** The time that the operation completed. */ |
| 2359 core.String endTime; | 2507 core.String endTime; |
| 2360 /** A message containing the detailed operation state. */ | 2508 /** A message containing the detailed operation state. */ |
| 2361 core.String innerState; | 2509 core.String innerState; |
| 2362 /** The time that the operation was requested. */ | 2510 /** The time that the operation was requested. */ |
| 2363 core.String insertTime; | 2511 core.String insertTime; |
| 2364 /** [Output-only] The operation type. */ | 2512 /** Output-only The operation type. */ |
| 2365 core.String operationType; | 2513 core.String operationType; |
| 2366 /** The time that the operation was started by the server. */ | 2514 /** The time that the operation was started by the server. */ |
| 2367 core.String startTime; | 2515 core.String startTime; |
| 2368 /** | 2516 /** |
| 2369 * A message containing the operation state. | 2517 * A message containing the operation state. |
| 2370 * Possible string values are: | 2518 * Possible string values are: |
| 2371 * - "UNKNOWN" : A UNKNOWN. | 2519 * - "UNKNOWN" : Unused. |
| 2372 * - "PENDING" : A PENDING. | 2520 * - "PENDING" : The operation has been created. |
| 2373 * - "RUNNING" : A RUNNING. | 2521 * - "RUNNING" : The operation is currently running. |
| 2374 * - "DONE" : A DONE. | 2522 * - "DONE" : The operation is done, either cancelled or completed. |
| 2375 */ | 2523 */ |
| 2376 core.String state; | 2524 core.String state; |
| 2377 /** [Output-only] Current operation status. */ | 2525 /** Output-only Current operation status. */ |
| 2378 OperationStatus status; | 2526 OperationStatus status; |
| 2379 /** [Output-only] Previous operation status. */ | 2527 /** Output-only Previous operation status. */ |
| 2380 core.List<OperationStatus> statusHistory; | 2528 core.List<OperationStatus> statusHistory; |
| 2529 /** Output-only Errors encountered during operation execution. */ |
| 2530 core.List<core.String> warnings; |
| 2381 | 2531 |
| 2382 OperationMetadata(); | 2532 OperationMetadata(); |
| 2383 | 2533 |
| 2384 OperationMetadata.fromJson(core.Map _json) { | 2534 OperationMetadata.fromJson(core.Map _json) { |
| 2385 if (_json.containsKey("clusterName")) { | 2535 if (_json.containsKey("clusterName")) { |
| 2386 clusterName = _json["clusterName"]; | 2536 clusterName = _json["clusterName"]; |
| 2387 } | 2537 } |
| 2388 if (_json.containsKey("clusterUuid")) { | 2538 if (_json.containsKey("clusterUuid")) { |
| 2389 clusterUuid = _json["clusterUuid"]; | 2539 clusterUuid = _json["clusterUuid"]; |
| 2390 } | 2540 } |
| (...skipping 20 matching lines...) Expand all Loading... |
| 2411 } | 2561 } |
| 2412 if (_json.containsKey("state")) { | 2562 if (_json.containsKey("state")) { |
| 2413 state = _json["state"]; | 2563 state = _json["state"]; |
| 2414 } | 2564 } |
| 2415 if (_json.containsKey("status")) { | 2565 if (_json.containsKey("status")) { |
| 2416 status = new OperationStatus.fromJson(_json["status"]); | 2566 status = new OperationStatus.fromJson(_json["status"]); |
| 2417 } | 2567 } |
| 2418 if (_json.containsKey("statusHistory")) { | 2568 if (_json.containsKey("statusHistory")) { |
| 2419 statusHistory = _json["statusHistory"].map((value) => new OperationStatus.
fromJson(value)).toList(); | 2569 statusHistory = _json["statusHistory"].map((value) => new OperationStatus.
fromJson(value)).toList(); |
| 2420 } | 2570 } |
| 2571 if (_json.containsKey("warnings")) { |
| 2572 warnings = _json["warnings"]; |
| 2573 } |
| 2421 } | 2574 } |
| 2422 | 2575 |
| 2423 core.Map toJson() { | 2576 core.Map toJson() { |
| 2424 var _json = new core.Map(); | 2577 var _json = new core.Map(); |
| 2425 if (clusterName != null) { | 2578 if (clusterName != null) { |
| 2426 _json["clusterName"] = clusterName; | 2579 _json["clusterName"] = clusterName; |
| 2427 } | 2580 } |
| 2428 if (clusterUuid != null) { | 2581 if (clusterUuid != null) { |
| 2429 _json["clusterUuid"] = clusterUuid; | 2582 _json["clusterUuid"] = clusterUuid; |
| 2430 } | 2583 } |
| (...skipping 20 matching lines...) Expand all Loading... |
| 2451 } | 2604 } |
| 2452 if (state != null) { | 2605 if (state != null) { |
| 2453 _json["state"] = state; | 2606 _json["state"] = state; |
| 2454 } | 2607 } |
| 2455 if (status != null) { | 2608 if (status != null) { |
| 2456 _json["status"] = (status).toJson(); | 2609 _json["status"] = (status).toJson(); |
| 2457 } | 2610 } |
| 2458 if (statusHistory != null) { | 2611 if (statusHistory != null) { |
| 2459 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); | 2612 _json["statusHistory"] = statusHistory.map((value) => (value).toJson()).to
List(); |
| 2460 } | 2613 } |
| 2614 if (warnings != null) { |
| 2615 _json["warnings"] = warnings; |
| 2616 } |
| 2461 return _json; | 2617 return _json; |
| 2462 } | 2618 } |
| 2463 } | 2619 } |
| 2464 | 2620 |
| 2465 /** The status of the operation. */ | 2621 /** The status of the operation. */ |
| 2466 class OperationStatus { | 2622 class OperationStatus { |
| 2467 /** A message containing any operation metadata details. */ | 2623 /** A message containing any operation metadata details. */ |
| 2468 core.String details; | 2624 core.String details; |
| 2469 /** A message containing the detailed operation state. */ | 2625 /** A message containing the detailed operation state. */ |
| 2470 core.String innerState; | 2626 core.String innerState; |
| 2471 /** | 2627 /** |
| 2472 * A message containing the operation state. | 2628 * A message containing the operation state. |
| 2473 * Possible string values are: | 2629 * Possible string values are: |
| 2474 * - "UNKNOWN" : A UNKNOWN. | 2630 * - "UNKNOWN" : Unused. |
| 2475 * - "PENDING" : A PENDING. | 2631 * - "PENDING" : The operation has been created. |
| 2476 * - "RUNNING" : A RUNNING. | 2632 * - "RUNNING" : The operation is running. |
| 2477 * - "DONE" : A DONE. | 2633 * - "DONE" : The operation is done; either cancelled or completed. |
| 2478 */ | 2634 */ |
| 2479 core.String state; | 2635 core.String state; |
| 2480 /** The time this state was entered. */ | 2636 /** The time this state was entered. */ |
| 2481 core.String stateStartTime; | 2637 core.String stateStartTime; |
| 2482 | 2638 |
| 2483 OperationStatus(); | 2639 OperationStatus(); |
| 2484 | 2640 |
| 2485 OperationStatus.fromJson(core.Map _json) { | 2641 OperationStatus.fromJson(core.Map _json) { |
| 2486 if (_json.containsKey("details")) { | 2642 if (_json.containsKey("details")) { |
| 2487 details = _json["details"]; | 2643 details = _json["details"]; |
| (...skipping 23 matching lines...) Expand all Loading... |
| 2511 if (stateStartTime != null) { | 2667 if (stateStartTime != null) { |
| 2512 _json["stateStartTime"] = stateStartTime; | 2668 _json["stateStartTime"] = stateStartTime; |
| 2513 } | 2669 } |
| 2514 return _json; | 2670 return _json; |
| 2515 } | 2671 } |
| 2516 } | 2672 } |
| 2517 | 2673 |
| 2518 /** A Cloud Dataproc job for running Pig queries on YARN. */ | 2674 /** A Cloud Dataproc job for running Pig queries on YARN. */ |
| 2519 class PigJob { | 2675 class PigJob { |
| 2520 /** | 2676 /** |
| 2521 * [Optional] Whether to continue executing queries if a query fails. The | 2677 * Optional Whether to continue executing queries if a query fails. The |
| 2522 * default value is `false`. Setting to `true` can be useful when executing | 2678 * default value is false. Setting to true can be useful when executing |
| 2523 * independent parallel queries. | 2679 * independent parallel queries. |
| 2524 */ | 2680 */ |
| 2525 core.bool continueOnFailure; | 2681 core.bool continueOnFailure; |
| 2526 /** | 2682 /** |
| 2527 * [Optional] HCFS URIs of jar files to add to the CLASSPATH of the Pig Client | 2683 * Optional HCFS URIs of jar files to add to the CLASSPATH of the Pig Client |
| 2528 * and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs. | 2684 * and Hadoop MapReduce (MR) tasks. Can contain Pig UDFs. |
| 2529 */ | 2685 */ |
| 2530 core.List<core.String> jarFileUris; | 2686 core.List<core.String> jarFileUris; |
| 2531 /** [Optional] The runtime log configuration for job execution. */ | 2687 /** Optional The runtime log configuration for job execution. */ |
| 2532 LoggingConfiguration loggingConfiguration; | 2688 LoggingConfiguration loggingConfiguration; |
| 2533 /** | 2689 /** |
| 2534 * [Optional] A mapping of property names to values, used to configure Pig. | 2690 * Optional A mapping of property names to values, used to configure Pig. |
| 2535 * Properties that conflict with values set by the Cloud Dataproc API may be | 2691 * Properties that conflict with values set by the Cloud Dataproc API may be |
| 2536 * overwritten. Can include properties set in /etc/hadoop/conf / * -site.xml, | 2692 * overwritten. Can include properties set in /etc/hadoop/conf / * -site.xml, |
| 2537 * /etc/pig/conf/pig.properties, and classes in user code. | 2693 * /etc/pig/conf/pig.properties, and classes in user code. |
| 2538 */ | 2694 */ |
| 2539 core.Map<core.String, core.String> properties; | 2695 core.Map<core.String, core.String> properties; |
| 2540 /** The HCFS URI of the script that contains the Pig queries. */ | 2696 /** The HCFS URI of the script that contains the Pig queries. */ |
| 2541 core.String queryFileUri; | 2697 core.String queryFileUri; |
| 2542 /** A list of queries. */ | 2698 /** A list of queries. */ |
| 2543 QueryList queryList; | 2699 QueryList queryList; |
| 2544 /** | 2700 /** |
| 2545 * [Optional] Mapping of query variable names to values (equivalent to the Pig | 2701 * Optional Mapping of query variable names to values (equivalent to the Pig |
| 2546 * command: `name=[value]`). | 2702 * command: name=[value]). |
| 2547 */ | 2703 */ |
| 2548 core.Map<core.String, core.String> scriptVariables; | 2704 core.Map<core.String, core.String> scriptVariables; |
| 2549 | 2705 |
| 2550 PigJob(); | 2706 PigJob(); |
| 2551 | 2707 |
| 2552 PigJob.fromJson(core.Map _json) { | 2708 PigJob.fromJson(core.Map _json) { |
| 2553 if (_json.containsKey("continueOnFailure")) { | 2709 if (_json.containsKey("continueOnFailure")) { |
| 2554 continueOnFailure = _json["continueOnFailure"]; | 2710 continueOnFailure = _json["continueOnFailure"]; |
| 2555 } | 2711 } |
| 2556 if (_json.containsKey("jarFileUris")) { | 2712 if (_json.containsKey("jarFileUris")) { |
| (...skipping 39 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2596 if (scriptVariables != null) { | 2752 if (scriptVariables != null) { |
| 2597 _json["scriptVariables"] = scriptVariables; | 2753 _json["scriptVariables"] = scriptVariables; |
| 2598 } | 2754 } |
| 2599 return _json; | 2755 return _json; |
| 2600 } | 2756 } |
| 2601 } | 2757 } |
| 2602 | 2758 |
| 2603 /** A Cloud Dataproc job for running PySpark applications on YARN. */ | 2759 /** A Cloud Dataproc job for running PySpark applications on YARN. */ |
| 2604 class PySparkJob { | 2760 class PySparkJob { |
| 2605 /** | 2761 /** |
| 2606 * [Optional] HCFS URIs of archives to be extracted in the working directory | 2762 * Optional HCFS URIs of archives to be extracted in the working directory of |
| 2607 * of .jar, .tar, .tar.gz, .tgz, and .zip. | 2763 * .jar, .tar, .tar.gz, .tgz, and .zip. |
| 2608 */ | 2764 */ |
| 2609 core.List<core.String> archiveUris; | 2765 core.List<core.String> archiveUris; |
| 2610 /** | 2766 /** |
| 2611 * [Optional] The arguments to pass to the driver. Do not include arguments, | 2767 * Optional The arguments to pass to the driver. Do not include arguments, |
| 2612 * such as `--conf`, that can be set as job properties, since a collision may | 2768 * such as --conf, that can be set as job properties, since a collision may |
| 2613 * occur that causes an incorrect job submission. | 2769 * occur that causes an incorrect job submission. |
| 2614 */ | 2770 */ |
| 2615 core.List<core.String> args; | 2771 core.List<core.String> args; |
| 2616 /** | 2772 /** |
| 2617 * [Optional] HCFS URIs of files to be copied to the working directory of | 2773 * Optional HCFS URIs of files to be copied to the working directory of Python |
| 2618 * Python drivers and distributed tasks. Useful for naively parallel tasks. | 2774 * drivers and distributed tasks. Useful for naively parallel tasks. |
| 2619 */ | 2775 */ |
| 2620 core.List<core.String> fileUris; | 2776 core.List<core.String> fileUris; |
| 2621 /** | 2777 /** |
| 2622 * [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Python | 2778 * Optional HCFS URIs of jar files to add to the CLASSPATHs of the Python |
| 2623 * driver and tasks. | 2779 * driver and tasks. |
| 2624 */ | 2780 */ |
| 2625 core.List<core.String> jarFileUris; | 2781 core.List<core.String> jarFileUris; |
| 2626 /** [Optional] The runtime log configuration for job execution. */ | 2782 /** Optional The runtime log configuration for job execution. */ |
| 2627 LoggingConfiguration loggingConfiguration; | 2783 LoggingConfiguration loggingConfiguration; |
| 2628 /** | 2784 /** |
| 2629 * [Required] The Hadoop Compatible Filesystem (HCFS) URI of the main Python | 2785 * Required The Hadoop Compatible Filesystem (HCFS) URI of the main Python |
| 2630 * file to use as the driver. Must be a .py file. | 2786 * file to use as the driver. Must be a .py file. |
| 2631 */ | 2787 */ |
| 2632 core.String mainPythonFileUri; | 2788 core.String mainPythonFileUri; |
| 2633 /** | 2789 /** |
| 2634 * [Optional] A mapping of property names to values, used to configure | 2790 * Optional A mapping of property names to values, used to configure PySpark. |
| 2635 * PySpark. Properties that conflict with values set by the Cloud Dataproc API | 2791 * Properties that conflict with values set by the Cloud Dataproc API may be |
| 2636 * may be overwritten. Can include properties set in | 2792 * overwritten. Can include properties set in |
| 2637 * /etc/spark/conf/spark-defaults.conf and classes in user code. | 2793 * /etc/spark/conf/spark-defaults.conf and classes in user code. |
| 2638 */ | 2794 */ |
| 2639 core.Map<core.String, core.String> properties; | 2795 core.Map<core.String, core.String> properties; |
| 2640 /** | 2796 /** |
| 2641 * [Optional] HCFS file URIs of Python files to pass to the PySpark framework. | 2797 * Optional HCFS file URIs of Python files to pass to the PySpark framework. |
| 2642 * Supported file types: .py, .egg, and .zip. | 2798 * Supported file types: .py, .egg, and .zip. |
| 2643 */ | 2799 */ |
| 2644 core.List<core.String> pythonFileUris; | 2800 core.List<core.String> pythonFileUris; |
| 2645 | 2801 |
| 2646 PySparkJob(); | 2802 PySparkJob(); |
| 2647 | 2803 |
| 2648 PySparkJob.fromJson(core.Map _json) { | 2804 PySparkJob.fromJson(core.Map _json) { |
| 2649 if (_json.containsKey("archiveUris")) { | 2805 if (_json.containsKey("archiveUris")) { |
| 2650 archiveUris = _json["archiveUris"]; | 2806 archiveUris = _json["archiveUris"]; |
| 2651 } | 2807 } |
| (...skipping 46 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2698 if (pythonFileUris != null) { | 2854 if (pythonFileUris != null) { |
| 2699 _json["pythonFileUris"] = pythonFileUris; | 2855 _json["pythonFileUris"] = pythonFileUris; |
| 2700 } | 2856 } |
| 2701 return _json; | 2857 return _json; |
| 2702 } | 2858 } |
| 2703 } | 2859 } |
| 2704 | 2860 |
| 2705 /** A list of queries to run on a cluster. */ | 2861 /** A list of queries to run on a cluster. */ |
| 2706 class QueryList { | 2862 class QueryList { |
| 2707 /** | 2863 /** |
| 2708 * [Required] The queries to execute. You do not need to terminate a query | 2864 * Required The queries to execute. You do not need to terminate a query with |
| 2709 * with a semicolon. Multiple queries can be specified in one string by | 2865 * a semicolon. Multiple queries can be specified in one string by separating |
| 2710 * separating each with a semicolon. Here is an example of an Cloud Dataproc | 2866 * each with a semicolon. Here is an example of an Cloud Dataproc API snippet |
| 2711 * API snippet that uses a QueryList to specify a HiveJob: "hiveJob": { | 2867 * that uses a QueryList to specify a HiveJob: |
| 2712 * "queryList": { "queries": [ "query1", "query2", "query3;query4", ] } } | 2868 * "hiveJob": { |
| 2869 * "queryList": { |
| 2870 * "queries": [ |
| 2871 * "query1", |
| 2872 * "query2", |
| 2873 * "query3;query4", |
| 2874 * ] |
| 2875 * } |
| 2876 * } |
| 2713 */ | 2877 */ |
| 2714 core.List<core.String> queries; | 2878 core.List<core.String> queries; |
| 2715 | 2879 |
| 2716 QueryList(); | 2880 QueryList(); |
| 2717 | 2881 |
| 2718 QueryList.fromJson(core.Map _json) { | 2882 QueryList.fromJson(core.Map _json) { |
| 2719 if (_json.containsKey("queries")) { | 2883 if (_json.containsKey("queries")) { |
| 2720 queries = _json["queries"]; | 2884 queries = _json["queries"]; |
| 2721 } | 2885 } |
| 2722 } | 2886 } |
| 2723 | 2887 |
| 2724 core.Map toJson() { | 2888 core.Map toJson() { |
| 2725 var _json = new core.Map(); | 2889 var _json = new core.Map(); |
| 2726 if (queries != null) { | 2890 if (queries != null) { |
| 2727 _json["queries"] = queries; | 2891 _json["queries"] = queries; |
| 2728 } | 2892 } |
| 2729 return _json; | 2893 return _json; |
| 2730 } | 2894 } |
| 2731 } | 2895 } |
| 2732 | 2896 |
| 2733 /** | 2897 /** |
| 2734 * Specifies the selection and configuration of software inside the cluster. | 2898 * Specifies the selection and configuration of software inside the cluster. |
| 2735 */ | 2899 */ |
| 2736 class SoftwareConfiguration { | 2900 class SoftwareConfiguration { |
| 2737 /** | 2901 /** |
| 2738 * [Optional] The version of software inside the cluster. It must match the | 2902 * Optional The version of software inside the cluster. It must match the |
| 2739 * regular expression `[0-9]+\.[0-9]+`. If unspecified, it defaults to the | 2903 * regular expression [0-9]+\.[0-9]+. If unspecified, it defaults to the |
| 2740 * latest version (see [Cloud Dataproc Versioning](/dataproc/versioning)). | 2904 * latest version (see Cloud Dataproc Versioning). |
| 2741 */ | 2905 */ |
| 2742 core.String imageVersion; | 2906 core.String imageVersion; |
| 2743 /** | 2907 /** |
| 2744 * [Optional] The properties to set on daemon configuration files. Property | 2908 * Optional The properties to set on daemon configuration files.Property keys |
| 2745 * keys are specified in "prefix:property" format, such as | 2909 * are specified in "prefix:property" format, such as "core:fs.defaultFS". The |
| 2746 * "core:fs.defaultFS". The following are supported prefixes and their | 2910 * following are supported prefixes and their mappings: core - core-site.xml |
| 2747 * mappings: core - core-site.xml hdfs - hdfs-site.xml mapred - | 2911 * hdfs - hdfs-site.xml mapred - mapred-site.xml yarn - yarn-site.xml hive |
| 2748 * mapred-site.xml yarn - yarn-site.xml hive - hive-site.xml pig - | 2912 * - hive-site.xml pig - pig.properties spark - spark-defaults.conf |
| 2749 * pig.properties spark - spark-defaults.conf | |
| 2750 */ | 2913 */ |
| 2751 core.Map<core.String, core.String> properties; | 2914 core.Map<core.String, core.String> properties; |
| 2752 | 2915 |
| 2753 SoftwareConfiguration(); | 2916 SoftwareConfiguration(); |
| 2754 | 2917 |
| 2755 SoftwareConfiguration.fromJson(core.Map _json) { | 2918 SoftwareConfiguration.fromJson(core.Map _json) { |
| 2756 if (_json.containsKey("imageVersion")) { | 2919 if (_json.containsKey("imageVersion")) { |
| 2757 imageVersion = _json["imageVersion"]; | 2920 imageVersion = _json["imageVersion"]; |
| 2758 } | 2921 } |
| 2759 if (_json.containsKey("properties")) { | 2922 if (_json.containsKey("properties")) { |
| 2760 properties = _json["properties"]; | 2923 properties = _json["properties"]; |
| 2761 } | 2924 } |
| 2762 } | 2925 } |
| 2763 | 2926 |
| 2764 core.Map toJson() { | 2927 core.Map toJson() { |
| 2765 var _json = new core.Map(); | 2928 var _json = new core.Map(); |
| 2766 if (imageVersion != null) { | 2929 if (imageVersion != null) { |
| 2767 _json["imageVersion"] = imageVersion; | 2930 _json["imageVersion"] = imageVersion; |
| 2768 } | 2931 } |
| 2769 if (properties != null) { | 2932 if (properties != null) { |
| 2770 _json["properties"] = properties; | 2933 _json["properties"] = properties; |
| 2771 } | 2934 } |
| 2772 return _json; | 2935 return _json; |
| 2773 } | 2936 } |
| 2774 } | 2937 } |
| 2775 | 2938 |
| 2776 /** A Cloud Dataproc job for running Spark applications on YARN. */ | 2939 /** A Cloud Dataproc job for running Spark applications on YARN. */ |
| 2777 class SparkJob { | 2940 class SparkJob { |
| 2778 /** | 2941 /** |
| 2779 * [Optional] HCFS URIs of archives to be extracted in the working directory | 2942 * Optional HCFS URIs of archives to be extracted in the working directory of |
| 2780 * of Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, | 2943 * Spark drivers and tasks. Supported file types: .jar, .tar, .tar.gz, .tgz, |
| 2781 * .tgz, and .zip. | 2944 * and .zip. |
| 2782 */ | 2945 */ |
| 2783 core.List<core.String> archiveUris; | 2946 core.List<core.String> archiveUris; |
| 2784 /** | 2947 /** |
| 2785 * [Optional] The arguments to pass to the driver. Do not include arguments, | 2948 * Optional The arguments to pass to the driver. Do not include arguments, |
| 2786 * such as `--conf`, that can be set as job properties, since a collision may | 2949 * such as --conf, that can be set as job properties, since a collision may |
| 2787 * occur that causes an incorrect job submission. | 2950 * occur that causes an incorrect job submission. |
| 2788 */ | 2951 */ |
| 2789 core.List<core.String> args; | 2952 core.List<core.String> args; |
| 2790 /** | 2953 /** |
| 2791 * [Optional] HCFS URIs of files to be copied to the working directory of | 2954 * Optional HCFS URIs of files to be copied to the working directory of Spark |
| 2792 * Spark drivers and distributed tasks. Useful for naively parallel tasks. | 2955 * drivers and distributed tasks. Useful for naively parallel tasks. |
| 2793 */ | 2956 */ |
| 2794 core.List<core.String> fileUris; | 2957 core.List<core.String> fileUris; |
| 2795 /** | 2958 /** |
| 2796 * [Optional] HCFS URIs of jar files to add to the CLASSPATHs of the Spark | 2959 * Optional HCFS URIs of jar files to add to the CLASSPATHs of the Spark |
| 2797 * driver and tasks. | 2960 * driver and tasks. |
| 2798 */ | 2961 */ |
| 2799 core.List<core.String> jarFileUris; | 2962 core.List<core.String> jarFileUris; |
| 2800 /** [Optional] The runtime log configuration for job execution. */ | 2963 /** Optional The runtime log configuration for job execution. */ |
| 2801 LoggingConfiguration loggingConfiguration; | 2964 LoggingConfiguration loggingConfiguration; |
| 2802 /** | 2965 /** |
| 2803 * The name of the driver's main class. The jar file that contains the class | 2966 * The name of the driver's main class. The jar file that contains the class |
| 2804 * must be in the default CLASSPATH or specified in `jar_file_uris`. | 2967 * must be in the default CLASSPATH or specified in jar_file_uris. |
| 2805 */ | 2968 */ |
| 2806 core.String mainClass; | 2969 core.String mainClass; |
| 2807 /** | 2970 /** |
| 2808 * The Hadoop Compatible Filesystem (HCFS) URI of the jar file that contains | 2971 * The Hadoop Compatible Filesystem (HCFS) URI of the jar file that contains |
| 2809 * the main class. | 2972 * the main class. |
| 2810 */ | 2973 */ |
| 2811 core.String mainJarFileUri; | 2974 core.String mainJarFileUri; |
| 2812 /** | 2975 /** |
| 2813 * [Optional] A mapping of property names to values, used to configure Spark. | 2976 * Optional A mapping of property names to values, used to configure Spark. |
| 2814 * Properties that conflict with values set by the Cloud Dataproc API may be | 2977 * Properties that conflict with values set by the Cloud Dataproc API may be |
| 2815 * overwritten. Can include properties set in | 2978 * overwritten. Can include properties set in |
| 2816 * /etc/spark/conf/spark-defaults.conf and classes in user code. | 2979 * /etc/spark/conf/spark-defaults.conf and classes in user code. |
| 2817 */ | 2980 */ |
| 2818 core.Map<core.String, core.String> properties; | 2981 core.Map<core.String, core.String> properties; |
| 2819 | 2982 |
| 2820 SparkJob(); | 2983 SparkJob(); |
| 2821 | 2984 |
| 2822 SparkJob.fromJson(core.Map _json) { | 2985 SparkJob.fromJson(core.Map _json) { |
| 2823 if (_json.containsKey("archiveUris")) { | 2986 if (_json.containsKey("archiveUris")) { |
| (...skipping 47 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2871 } | 3034 } |
| 2872 if (properties != null) { | 3035 if (properties != null) { |
| 2873 _json["properties"] = properties; | 3036 _json["properties"] = properties; |
| 2874 } | 3037 } |
| 2875 return _json; | 3038 return _json; |
| 2876 } | 3039 } |
| 2877 } | 3040 } |
| 2878 | 3041 |
| 2879 /** A Cloud Dataproc job for running Spark SQL queries. */ | 3042 /** A Cloud Dataproc job for running Spark SQL queries. */ |
| 2880 class SparkSqlJob { | 3043 class SparkSqlJob { |
| 2881 /** [Optional] HCFS URIs of jar files to be added to the Spark CLASSPATH. */ | 3044 /** Optional HCFS URIs of jar files to be added to the Spark CLASSPATH. */ |
| 2882 core.List<core.String> jarFileUris; | 3045 core.List<core.String> jarFileUris; |
| 2883 /** [Optional] The runtime log configuration for job execution. */ | 3046 /** Optional The runtime log configuration for job execution. */ |
| 2884 LoggingConfiguration loggingConfiguration; | 3047 LoggingConfiguration loggingConfiguration; |
| 2885 /** | 3048 /** |
| 2886 * [Optional] A mapping of property names to values, used to configure Spark | 3049 * Optional A mapping of property names to values, used to configure Spark |
| 2887 * SQL's SparkConf. Properties that conflict with values set by the Cloud | 3050 * SQL's SparkConf. Properties that conflict with values set by the Cloud |
| 2888 * Dataproc API may be overwritten. | 3051 * Dataproc API may be overwritten. |
| 2889 */ | 3052 */ |
| 2890 core.Map<core.String, core.String> properties; | 3053 core.Map<core.String, core.String> properties; |
| 2891 /** The HCFS URI of the script that contains SQL queries. */ | 3054 /** The HCFS URI of the script that contains SQL queries. */ |
| 2892 core.String queryFileUri; | 3055 core.String queryFileUri; |
| 2893 /** A list of queries. */ | 3056 /** A list of queries. */ |
| 2894 QueryList queryList; | 3057 QueryList queryList; |
| 2895 /** | 3058 /** |
| 2896 * [Optional] Mapping of query variable names to values (equivalent to the | 3059 * Optional Mapping of query variable names to values (equivalent to the Spark |
| 2897 * Spark SQL command: SET `name="value";`). | 3060 * SQL command: SET name="value";). |
| 2898 */ | 3061 */ |
| 2899 core.Map<core.String, core.String> scriptVariables; | 3062 core.Map<core.String, core.String> scriptVariables; |
| 2900 | 3063 |
| 2901 SparkSqlJob(); | 3064 SparkSqlJob(); |
| 2902 | 3065 |
| 2903 SparkSqlJob.fromJson(core.Map _json) { | 3066 SparkSqlJob.fromJson(core.Map _json) { |
| 2904 if (_json.containsKey("jarFileUris")) { | 3067 if (_json.containsKey("jarFileUris")) { |
| 2905 jarFileUris = _json["jarFileUris"]; | 3068 jarFileUris = _json["jarFileUris"]; |
| 2906 } | 3069 } |
| 2907 if (_json.containsKey("loggingConfiguration")) { | 3070 if (_json.containsKey("loggingConfiguration")) { |
| (...skipping 31 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
| 2939 _json["queryList"] = (queryList).toJson(); | 3102 _json["queryList"] = (queryList).toJson(); |
| 2940 } | 3103 } |
| 2941 if (scriptVariables != null) { | 3104 if (scriptVariables != null) { |
| 2942 _json["scriptVariables"] = scriptVariables; | 3105 _json["scriptVariables"] = scriptVariables; |
| 2943 } | 3106 } |
| 2944 return _json; | 3107 return _json; |
| 2945 } | 3108 } |
| 2946 } | 3109 } |
| 2947 | 3110 |
| 2948 /** | 3111 /** |
| 2949 * The `Status` type defines a logical error model that is suitable for | 3112 * The Status type defines a logical error model that is suitable for different |
| 2950 * different programming environments, including REST APIs and RPC APIs. It is | 3113 * programming environments, including REST APIs and RPC APIs. It is used by |
| 2951 * used by [gRPC](https://github.com/grpc). The error model is designed to be: - | 3114 * gRPC (https://github.com/grpc). The error model is designed to be: |
| 2952 * Simple to use and understand for most users - Flexible enough to meet | 3115 * Simple to use and understand for most users |
| 2953 * unexpected needs # Overview The `Status` message contains three pieces of | 3116 * Flexible enough to meet unexpected needsOverviewThe Status message contains |
| 2954 * data: error code, error message, and error details. The error code should be | 3117 * three pieces of data: error code, error message, and error details. The error |
| 2955 * an enum value of google.rpc.Code, but it may accept additional error codes if | 3118 * code should be an enum value of google.rpc.Code, but it may accept additional |
| 2956 * needed. The error message should be a developer-facing English message that | 3119 * error codes if needed. The error message should be a developer-facing English |
| 2957 * helps developers *understand* and *resolve* the error. If a localized | 3120 * message that helps developers understand and resolve the error. If a |
| 2958 * user-facing error message is needed, put the localized message in the error | 3121 * localized user-facing error message is needed, put the localized message in |
| 2959 * details or localize it in the client. The optional error details may contain | 3122 * the error details or localize it in the client. The optional error details |
| 2960 * arbitrary information about the error. There is a predefined set of error | 3123 * may contain arbitrary information about the error. There is a predefined set |
| 2961 * detail types in the package `google.rpc` which can be used for common error | 3124 * of error detail types in the package google.rpc which can be used for common |
| 2962 * conditions. # Language mapping The `Status` message is the logical | 3125 * error conditions.Language mappingThe Status message is the logical |
| 2963 * representation of the error model, but it is not necessarily the actual wire | 3126 * representation of the error model, but it is not necessarily the actual wire |
| 2964 * format. When the `Status` message is exposed in different client libraries | 3127 * format. When the Status message is exposed in different client libraries and |
| 2965 * and different wire protocols, it can be mapped differently. For example, it | 3128 * different wire protocols, it can be mapped differently. For example, it will |
| 2966 * will likely be mapped to some exceptions in Java, but more likely mapped to | 3129 * likely be mapped to some exceptions in Java, but more likely mapped to some |
| 2967 * some error codes in C. # Other uses The error model and the `Status` message | 3130 * error codes in C.Other usesThe error model and the Status message can be used |
| 2968 * can be used in a variety of environments, either with or without APIs, to | 3131 * in a variety of environments, either with or without APIs, to provide a |
| 2969 * provide a consistent developer experience across different environments. | 3132 * consistent developer experience across different environments.Example uses of |
| 2970 * Example uses of this error model include: - Partial errors. If a service | 3133 * this error model include: |
| 2971 * needs to return partial errors to the client, it may embed the `Status` in | 3134 * Partial errors. If a service needs to return partial errors to the client, it |
| 2972 * the normal response to indicate the partial errors. - Workflow errors. A | 3135 * may embed the Status in the normal response to indicate the partial errors. |
| 2973 * typical workflow has multiple steps. Each step may have a `Status` message | 3136 * Workflow errors. A typical workflow has multiple steps. Each step may have a |
| 2974 * for error reporting purpose. - Batch operations. If a client uses batch | 3137 * Status message for error reporting purpose. |
| 2975 * request and batch response, the `Status` message should be used directly | 3138 * Batch operations. If a client uses batch request and batch response, the |
| 2976 * inside batch response, one for each error sub-response. - Asynchronous | 3139 * Status message should be used directly inside batch response, one for each |
| 2977 * operations. If an API call embeds asynchronous operation results in its | 3140 * error sub-response. |
| 2978 * response, the status of those operations should be represented directly using | 3141 * Asynchronous operations. If an API call embeds asynchronous operation results |
| 2979 * the `Status` message. - Logging. If some API errors are stored in logs, the | 3142 * in its response, the status of those operations should be represented |
| 2980 * message `Status` could be used directly after any stripping needed for | 3143 * directly using the Status message. |
| 2981 * security/privacy reasons. | 3144 * Logging. If some API errors are stored in logs, the message Status could be |
| 3145 * used directly after any stripping needed for security/privacy reasons. |
| 2982 */ | 3146 */ |
| 2983 class Status { | 3147 class Status { |
| 2984 /** The status code, which should be an enum value of google.rpc.Code. */ | 3148 /** The status code, which should be an enum value of google.rpc.Code. */ |
| 2985 core.int code; | 3149 core.int code; |
| 2986 /** | 3150 /** |
| 2987 * A list of messages that carry the error details. There will be a common set | 3151 * A list of messages that carry the error details. There will be a common set |
| 2988 * of message types for APIs to use. | 3152 * of message types for APIs to use. |
| 2989 * | 3153 * |
| 2990 * The values for Object must be JSON objects. It can consist of `num`, | 3154 * The values for Object must be JSON objects. It can consist of `num`, |
| 2991 * `String`, `bool` and `null` as well as `Map` and `List` values. | 3155 * `String`, `bool` and `null` as well as `Map` and `List` values. |
| (...skipping 30 matching lines...) Expand all Loading... |
| 3022 } | 3186 } |
| 3023 if (message != null) { | 3187 if (message != null) { |
| 3024 _json["message"] = message; | 3188 _json["message"] = message; |
| 3025 } | 3189 } |
| 3026 return _json; | 3190 return _json; |
| 3027 } | 3191 } |
| 3028 } | 3192 } |
| 3029 | 3193 |
| 3030 /** A request to submit a job. */ | 3194 /** A request to submit a job. */ |
| 3031 class SubmitJobRequest { | 3195 class SubmitJobRequest { |
| 3032 /** [Required] The job resource. */ | 3196 /** Required The job resource. */ |
| 3033 Job job; | 3197 Job job; |
| 3034 | 3198 |
| 3035 SubmitJobRequest(); | 3199 SubmitJobRequest(); |
| 3036 | 3200 |
| 3037 SubmitJobRequest.fromJson(core.Map _json) { | 3201 SubmitJobRequest.fromJson(core.Map _json) { |
| 3038 if (_json.containsKey("job")) { | 3202 if (_json.containsKey("job")) { |
| 3039 job = new Job.fromJson(_json["job"]); | 3203 job = new Job.fromJson(_json["job"]); |
| 3040 } | 3204 } |
| 3041 } | 3205 } |
| 3042 | 3206 |
| 3043 core.Map toJson() { | 3207 core.Map toJson() { |
| 3044 var _json = new core.Map(); | 3208 var _json = new core.Map(); |
| 3045 if (job != null) { | 3209 if (job != null) { |
| 3046 _json["job"] = (job).toJson(); | 3210 _json["job"] = (job).toJson(); |
| 3047 } | 3211 } |
| 3048 return _json; | 3212 return _json; |
| 3049 } | 3213 } |
| 3050 } | 3214 } |
| 3051 | 3215 |
| 3052 /** | 3216 /** |
| 3053 * A YARN application created by a job. Application information is a subset of | 3217 * A YARN application created by a job. Application information is a subset of |
| 3054 * org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto. | 3218 * <code>org.apache.hadoop.yarn.proto.YarnProtos.ApplicationReportProto</code>. |
| 3055 */ | 3219 */ |
| 3056 class YarnApplication { | 3220 class YarnApplication { |
| 3057 /** [Required] The application name. */ | 3221 /** Required The application name. */ |
| 3058 core.String name; | 3222 core.String name; |
| 3059 /** [Required] The numerical progress of the application, from 1 to 100. */ | 3223 /** Required The numerical progress of the application, from 1 to 100. */ |
| 3060 core.double progress; | 3224 core.double progress; |
| 3061 /** | 3225 /** |
| 3062 * [Required] The application state. | 3226 * Required The application state. |
| 3063 * Possible string values are: | 3227 * Possible string values are: |
| 3064 * - "STATE_UNSPECIFIED" : A STATE_UNSPECIFIED. | 3228 * - "STATE_UNSPECIFIED" : Status is unspecified. |
| 3065 * - "NEW" : A NEW. | 3229 * - "NEW" : Status is NEW. |
| 3066 * - "NEW_SAVING" : A NEW_SAVING. | 3230 * - "NEW_SAVING" : Status is NEW_SAVING. |
| 3067 * - "SUBMITTED" : A SUBMITTED. | 3231 * - "SUBMITTED" : Status is SUBMITTED. |
| 3068 * - "ACCEPTED" : A ACCEPTED. | 3232 * - "ACCEPTED" : Status is ACCEPTED. |
| 3069 * - "RUNNING" : A RUNNING. | 3233 * - "RUNNING" : Status is RUNNING. |
| 3070 * - "FINISHED" : A FINISHED. | 3234 * - "FINISHED" : Status is FINISHED. |
| 3071 * - "FAILED" : A FAILED. | 3235 * - "FAILED" : Status is FAILED. |
| 3072 * - "KILLED" : A KILLED. | 3236 * - "KILLED" : Status is KILLED. |
| 3073 */ | 3237 */ |
| 3074 core.String state; | 3238 core.String state; |
| 3075 /** | 3239 /** |
| 3076 * [Optional] The HTTP URL of the ApplicationMaster, HistoryServer, or | 3240 * Optional The HTTP URL of the ApplicationMaster, HistoryServer, or |
| 3077 * TimelineServer that provides application-specific information. The URL uses | 3241 * TimelineServer that provides application-specific information. The URL uses |
| 3078 * the internal hostname, and requires a proxy server for resolution and, | 3242 * the internal hostname, and requires a proxy server for resolution and, |
| 3079 * possibly, access. | 3243 * possibly, access. |
| 3080 */ | 3244 */ |
| 3081 core.String trackingUrl; | 3245 core.String trackingUrl; |
| 3082 | 3246 |
| 3083 YarnApplication(); | 3247 YarnApplication(); |
| 3084 | 3248 |
| 3085 YarnApplication.fromJson(core.Map _json) { | 3249 YarnApplication.fromJson(core.Map _json) { |
| 3086 if (_json.containsKey("name")) { | 3250 if (_json.containsKey("name")) { |
| (...skipping 20 matching lines...) Expand all Loading... |
| 3107 } | 3271 } |
| 3108 if (state != null) { | 3272 if (state != null) { |
| 3109 _json["state"] = state; | 3273 _json["state"] = state; |
| 3110 } | 3274 } |
| 3111 if (trackingUrl != null) { | 3275 if (trackingUrl != null) { |
| 3112 _json["trackingUrl"] = trackingUrl; | 3276 _json["trackingUrl"] = trackingUrl; |
| 3113 } | 3277 } |
| 3114 return _json; | 3278 return _json; |
| 3115 } | 3279 } |
| 3116 } | 3280 } |
| OLD | NEW |