Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(753)

Unified Diff: generated/googleapis_beta/lib/dataflow/v1b3.dart

Issue 3006323002: Api-Roll 54: 2017-09-11 (Closed)
Patch Set: use 2.0.0-dev.infinity sdk constraint in pubspecs Created 3 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: generated/googleapis_beta/lib/dataflow/v1b3.dart
diff --git a/generated/googleapis_beta/lib/dataflow/v1b3.dart b/generated/googleapis_beta/lib/dataflow/v1b3.dart
index 44c7280a4659b806997bbd1ac1be8c95061ba4f8..984143adab4704c490b22c6c7f03d7d922802ad6 100644
--- a/generated/googleapis_beta/lib/dataflow/v1b3.dart
+++ b/generated/googleapis_beta/lib/dataflow/v1b3.dart
@@ -9,63 +9,67 @@ import 'dart:convert' as convert;
import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons;
import 'package:http/http.dart' as http;
-export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show
- ApiRequestError, DetailedApiRequestError;
+export 'package:_discoveryapis_commons/_discoveryapis_commons.dart'
+ show ApiRequestError, DetailedApiRequestError;
const core.String USER_AGENT = 'dart-api-client dataflow/v1b3';
-/** Manages Google Cloud Dataflow projects on Google Cloud Platform. */
+/// Manages Google Cloud Dataflow projects on Google Cloud Platform.
class DataflowApi {
- /** View and manage your data across Google Cloud Platform services */
- static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platform";
+ /// View and manage your data across Google Cloud Platform services
+ static const CloudPlatformScope =
+ "https://www.googleapis.com/auth/cloud-platform";
- /** View and manage your Google Compute Engine resources */
+ /// View and manage your Google Compute Engine resources
static const ComputeScope = "https://www.googleapis.com/auth/compute";
- /** View your Google Compute Engine resources */
- static const ComputeReadonlyScope = "https://www.googleapis.com/auth/compute.readonly";
-
- /** View your email address */
- static const UserinfoEmailScope = "https://www.googleapis.com/auth/userinfo.email";
+ /// View your Google Compute Engine resources
+ static const ComputeReadonlyScope =
+ "https://www.googleapis.com/auth/compute.readonly";
+ /// View your email address
+ static const UserinfoEmailScope =
+ "https://www.googleapis.com/auth/userinfo.email";
final commons.ApiRequester _requester;
ProjectsResourceApi get projects => new ProjectsResourceApi(_requester);
- DataflowApi(http.Client client, {core.String rootUrl: "https://dataflow.googleapis.com/", core.String servicePath: ""}) :
- _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
+ DataflowApi(http.Client client,
+ {core.String rootUrl: "https://dataflow.googleapis.com/",
+ core.String servicePath: ""})
+ : _requester =
+ new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
}
-
class ProjectsResourceApi {
final commons.ApiRequester _requester;
ProjectsJobsResourceApi get jobs => new ProjectsJobsResourceApi(_requester);
- ProjectsLocationsResourceApi get locations => new ProjectsLocationsResourceApi(_requester);
- ProjectsTemplatesResourceApi get templates => new ProjectsTemplatesResourceApi(_requester);
-
- ProjectsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Send a worker_message to the service.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project to send the WorkerMessages to.
- *
- * Completes with a [SendWorkerMessagesResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<SendWorkerMessagesResponse> workerMessages(SendWorkerMessagesRequest request, core.String projectId) {
+ ProjectsLocationsResourceApi get locations =>
+ new ProjectsLocationsResourceApi(_requester);
+ ProjectsTemplatesResourceApi get templates =>
+ new ProjectsTemplatesResourceApi(_requester);
+
+ ProjectsResourceApi(commons.ApiRequester client) : _requester = client;
+
+ /// Send a worker_message to the service.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project to send the WorkerMessages to.
+ ///
+ /// Completes with a [SendWorkerMessagesResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<SendWorkerMessagesResponse> workerMessages(
+ SendWorkerMessagesRequest request, core.String projectId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -80,72 +84,77 @@ class ProjectsResourceApi {
throw new core.ArgumentError("Parameter projectId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/WorkerMessages';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/WorkerMessages';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new SendWorkerMessagesResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new SendWorkerMessagesResponse.fromJson(data));
}
-
}
-
class ProjectsJobsResourceApi {
final commons.ApiRequester _requester;
- ProjectsJobsDebugResourceApi get debug => new ProjectsJobsDebugResourceApi(_requester);
- ProjectsJobsMessagesResourceApi get messages => new ProjectsJobsMessagesResourceApi(_requester);
- ProjectsJobsWorkItemsResourceApi get workItems => new ProjectsJobsWorkItemsResourceApi(_requester);
-
- ProjectsJobsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * List the jobs of a project across all regions.
- *
- * Request parameters:
- *
- * [projectId] - The project which owns the jobs.
- *
- * [filter] - The kind of filter to use.
- * Possible string values are:
- * - "UNKNOWN" : A UNKNOWN.
- * - "ALL" : A ALL.
- * - "TERMINATED" : A TERMINATED.
- * - "ACTIVE" : A ACTIVE.
- *
- * [location] - The location that contains this job.
- *
- * [pageToken] - Set this to the 'next_page_token' field of a previous
- * response
- * to request additional results in a long list.
- *
- * [pageSize] - If there are many jobs, limit response to at most this many.
- * The actual number of jobs returned will be the lesser of max_responses
- * and an unspecified server-defined limit.
- *
- * [view] - Level of information requested in response. Default is
- * `JOB_VIEW_SUMMARY`.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [ListJobsResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListJobsResponse> aggregated(core.String projectId, {core.String filter, core.String location, core.String pageToken, core.int pageSize, core.String view}) {
+ ProjectsJobsDebugResourceApi get debug =>
+ new ProjectsJobsDebugResourceApi(_requester);
+ ProjectsJobsMessagesResourceApi get messages =>
+ new ProjectsJobsMessagesResourceApi(_requester);
+ ProjectsJobsWorkItemsResourceApi get workItems =>
+ new ProjectsJobsWorkItemsResourceApi(_requester);
+
+ ProjectsJobsResourceApi(commons.ApiRequester client) : _requester = client;
+
+ /// List the jobs of a project across all regions.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project which owns the jobs.
+ ///
+ /// [view] - Level of information requested in response. Default is
+ /// `JOB_VIEW_SUMMARY`.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// [filter] - The kind of filter to use.
+ /// Possible string values are:
+ /// - "UNKNOWN" : A UNKNOWN.
+ /// - "ALL" : A ALL.
+ /// - "TERMINATED" : A TERMINATED.
+ /// - "ACTIVE" : A ACTIVE.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [pageToken] - Set this to the 'next_page_token' field of a previous
+ /// response
+ /// to request additional results in a long list.
+ ///
+ /// [pageSize] - If there are many jobs, limit response to at most this many.
+ /// The actual number of jobs returned will be the lesser of max_responses
+ /// and an unspecified server-defined limit.
+ ///
+ /// Completes with a [ListJobsResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListJobsResponse> aggregated(core.String projectId,
+ {core.String view,
+ core.String filter,
+ core.String location,
+ core.String pageToken,
+ core.int pageSize}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -156,6 +165,9 @@ class ProjectsJobsResourceApi {
if (projectId == null) {
throw new core.ArgumentError("Parameter projectId is required.");
}
+ if (view != null) {
+ _queryParams["view"] = [view];
+ }
if (filter != null) {
_queryParams["filter"] = [filter];
}
@@ -168,51 +180,49 @@ class ProjectsJobsResourceApi {
if (pageSize != null) {
_queryParams["pageSize"] = ["${pageSize}"];
}
- if (view != null) {
- _queryParams["view"] = [view];
- }
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs:aggregated';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs:aggregated';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListJobsResponse.fromJson(data));
}
- /**
- * Creates a Cloud Dataflow job.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [location] - The location that contains this job.
- *
- * [replaceJobId] - Deprecated. This field is now in the Job message.
- *
- * [view] - The level of information requested in response.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> create(Job request, core.String projectId, {core.String location, core.String replaceJobId, core.String view}) {
+ /// Creates a Cloud Dataflow job.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [replaceJobId] - Deprecated. This field is now in the Job message.
+ ///
+ /// [view] - The level of information requested in response.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> create(Job request, core.String projectId,
+ {core.String location, core.String replaceJobId, core.String view}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -236,45 +246,46 @@ class ProjectsJobsResourceApi {
_queryParams["view"] = [view];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Gets the state of the specified Cloud Dataflow job.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [jobId] - The job ID.
- *
- * [location] - The location that contains this job.
- *
- * [view] - The level of information requested in response.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> get(core.String projectId, core.String jobId, {core.String location, core.String view}) {
+ /// Gets the state of the specified Cloud Dataflow job.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [jobId] - The job ID.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [view] - The level of information requested in response.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> get(core.String projectId, core.String jobId,
+ {core.String location, core.String view}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -295,41 +306,42 @@ class ProjectsJobsResourceApi {
_queryParams["view"] = [view];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId');
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId');
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Request the job status.
- *
- * Request parameters:
- *
- * [projectId] - A project id.
- *
- * [jobId] - The job to get messages for.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [startTime] - Return only metric data that has changed since this time.
- * Default is to return all information about all metrics for the job.
- *
- * Completes with a [JobMetrics].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<JobMetrics> getMetrics(core.String projectId, core.String jobId, {core.String location, core.String startTime}) {
+ /// Request the job status.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - A project id.
+ ///
+ /// [jobId] - The job to get messages for.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [startTime] - Return only metric data that has changed since this time.
+ /// Default is to return all information about all metrics for the job.
+ ///
+ /// Completes with a [JobMetrics].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<JobMetrics> getMetrics(core.String projectId, core.String jobId,
+ {core.String location, core.String startTime}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -350,59 +362,65 @@ class ProjectsJobsResourceApi {
_queryParams["startTime"] = [startTime];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/metrics';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/metrics';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new JobMetrics.fromJson(data));
}
- /**
- * List the jobs of a project in a given region.
- *
- * Request parameters:
- *
- * [projectId] - The project which owns the jobs.
- *
- * [view] - Level of information requested in response. Default is
- * `JOB_VIEW_SUMMARY`.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * [filter] - The kind of filter to use.
- * Possible string values are:
- * - "UNKNOWN" : A UNKNOWN.
- * - "ALL" : A ALL.
- * - "TERMINATED" : A TERMINATED.
- * - "ACTIVE" : A ACTIVE.
- *
- * [location] - The location that contains this job.
- *
- * [pageToken] - Set this to the 'next_page_token' field of a previous
- * response
- * to request additional results in a long list.
- *
- * [pageSize] - If there are many jobs, limit response to at most this many.
- * The actual number of jobs returned will be the lesser of max_responses
- * and an unspecified server-defined limit.
- *
- * Completes with a [ListJobsResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListJobsResponse> list(core.String projectId, {core.String view, core.String filter, core.String location, core.String pageToken, core.int pageSize}) {
+ /// List the jobs of a project in a given region.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project which owns the jobs.
+ ///
+ /// [filter] - The kind of filter to use.
+ /// Possible string values are:
+ /// - "UNKNOWN" : A UNKNOWN.
+ /// - "ALL" : A ALL.
+ /// - "TERMINATED" : A TERMINATED.
+ /// - "ACTIVE" : A ACTIVE.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [pageToken] - Set this to the 'next_page_token' field of a previous
+ /// response
+ /// to request additional results in a long list.
+ ///
+ /// [pageSize] - If there are many jobs, limit response to at most this many.
+ /// The actual number of jobs returned will be the lesser of max_responses
+ /// and an unspecified server-defined limit.
+ ///
+ /// [view] - Level of information requested in response. Default is
+ /// `JOB_VIEW_SUMMARY`.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// Completes with a [ListJobsResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListJobsResponse> list(core.String projectId,
+ {core.String filter,
+ core.String location,
+ core.String pageToken,
+ core.int pageSize,
+ core.String view}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -413,9 +431,6 @@ class ProjectsJobsResourceApi {
if (projectId == null) {
throw new core.ArgumentError("Parameter projectId is required.");
}
- if (view != null) {
- _queryParams["view"] = [view];
- }
if (filter != null) {
_queryParams["filter"] = [filter];
}
@@ -428,41 +443,46 @@ class ProjectsJobsResourceApi {
if (pageSize != null) {
_queryParams["pageSize"] = ["${pageSize}"];
}
+ if (view != null) {
+ _queryParams["view"] = [view];
+ }
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListJobsResponse.fromJson(data));
}
- /**
- * Updates the state of an existing Cloud Dataflow job.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [jobId] - The job ID.
- *
- * [location] - The location that contains this job.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> update(Job request, core.String projectId, core.String jobId, {core.String location}) {
+ /// Updates the state of an existing Cloud Dataflow job.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [jobId] - The job ID.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> update(
+ Job request, core.String projectId, core.String jobId,
+ {core.String location}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -483,47 +503,46 @@ class ProjectsJobsResourceApi {
_queryParams["location"] = [location];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId');
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId');
- var _response = _requester.request(_url,
- "PUT",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
-
}
-
class ProjectsJobsDebugResourceApi {
final commons.ApiRequester _requester;
- ProjectsJobsDebugResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Get encoded debug configuration for component. Not cacheable.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project id.
- *
- * [jobId] - The job id.
- *
- * Completes with a [GetDebugConfigResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<GetDebugConfigResponse> getConfig(GetDebugConfigRequest request, core.String projectId, core.String jobId) {
+ ProjectsJobsDebugResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Get encoded debug configuration for component. Not cacheable.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project id.
+ ///
+ /// [jobId] - The job id.
+ ///
+ /// Completes with a [GetDebugConfigResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<GetDebugConfigResponse> getConfig(
+ GetDebugConfigRequest request, core.String projectId, core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -541,38 +560,42 @@ class ProjectsJobsDebugResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/debug/getConfig';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/debug/getConfig';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new GetDebugConfigResponse.fromJson(data));
}
- /**
- * Send encoded debug capture data for component.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project id.
- *
- * [jobId] - The job id.
- *
- * Completes with a [SendDebugCaptureResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<SendDebugCaptureResponse> sendCapture(SendDebugCaptureRequest request, core.String projectId, core.String jobId) {
+ /// Send encoded debug capture data for component.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project id.
+ ///
+ /// [jobId] - The job id.
+ ///
+ /// Completes with a [SendDebugCaptureResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<SendDebugCaptureResponse> sendCapture(
+ SendDebugCaptureRequest request,
+ core.String projectId,
+ core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -590,73 +613,81 @@ class ProjectsJobsDebugResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/debug/sendCapture';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/debug/sendCapture';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new SendDebugCaptureResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new SendDebugCaptureResponse.fromJson(data));
}
-
}
-
class ProjectsJobsMessagesResourceApi {
final commons.ApiRequester _requester;
- ProjectsJobsMessagesResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Request the job status.
- *
- * Request parameters:
- *
- * [projectId] - A project id.
- *
- * [jobId] - The job to get messages about.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [endTime] - Return only messages with timestamps < end_time. The default is
- * now
- * (i.e. return up to the latest messages available).
- *
- * [startTime] - If specified, return only messages with timestamps >=
- * start_time.
- * The default is the job creation time (i.e. beginning of messages).
- *
- * [pageToken] - If supplied, this should be the value of next_page_token
- * returned
- * by an earlier call. This will cause the next page of results to
- * be returned.
- *
- * [pageSize] - If specified, determines the maximum number of messages to
- * return. If unspecified, the service may choose an appropriate
- * default, or may return an arbitrarily large number of results.
- *
- * [minimumImportance] - Filter to only get messages with importance >= level
- * Possible string values are:
- * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN.
- * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG.
- * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED.
- * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC.
- * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING.
- * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR.
- *
- * Completes with a [ListJobMessagesResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListJobMessagesResponse> list(core.String projectId, core.String jobId, {core.String location, core.String endTime, core.String startTime, core.String pageToken, core.int pageSize, core.String minimumImportance}) {
+ ProjectsJobsMessagesResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Request the job status.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - A project id.
+ ///
+ /// [jobId] - The job to get messages about.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [endTime] - Return only messages with timestamps < end_time. The default
+ /// is now
+ /// (i.e. return up to the latest messages available).
+ ///
+ /// [pageToken] - If supplied, this should be the value of next_page_token
+ /// returned
+ /// by an earlier call. This will cause the next page of results to
+ /// be returned.
+ ///
+ /// [startTime] - If specified, return only messages with timestamps >=
+ /// start_time.
+ /// The default is the job creation time (i.e. beginning of messages).
+ ///
+ /// [pageSize] - If specified, determines the maximum number of messages to
+ /// return. If unspecified, the service may choose an appropriate
+ /// default, or may return an arbitrarily large number of results.
+ ///
+ /// [minimumImportance] - Filter to only get messages with importance >=
+ /// level
+ /// Possible string values are:
+ /// - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN.
+ /// - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG.
+ /// - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED.
+ /// - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC.
+ /// - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING.
+ /// - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR.
+ ///
+ /// Completes with a [ListJobMessagesResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListJobMessagesResponse> list(
+ core.String projectId, core.String jobId,
+ {core.String location,
+ core.String endTime,
+ core.String pageToken,
+ core.String startTime,
+ core.int pageSize,
+ core.String minimumImportance}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -676,12 +707,12 @@ class ProjectsJobsMessagesResourceApi {
if (endTime != null) {
_queryParams["endTime"] = [endTime];
}
- if (startTime != null) {
- _queryParams["startTime"] = [startTime];
- }
if (pageToken != null) {
_queryParams["pageToken"] = [pageToken];
}
+ if (startTime != null) {
+ _queryParams["startTime"] = [startTime];
+ }
if (pageSize != null) {
_queryParams["pageSize"] = ["${pageSize}"];
}
@@ -689,47 +720,47 @@ class ProjectsJobsMessagesResourceApi {
_queryParams["minimumImportance"] = [minimumImportance];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/messages';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/messages';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListJobMessagesResponse.fromJson(data));
}
-
}
-
class ProjectsJobsWorkItemsResourceApi {
final commons.ApiRequester _requester;
- ProjectsJobsWorkItemsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Leases a dataflow WorkItem to run.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Identifies the project this worker belongs to.
- *
- * [jobId] - Identifies the workflow job this worker belongs to.
- *
- * Completes with a [LeaseWorkItemResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<LeaseWorkItemResponse> lease(LeaseWorkItemRequest request, core.String projectId, core.String jobId) {
+ ProjectsJobsWorkItemsResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Leases a dataflow WorkItem to run.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Identifies the project this worker belongs to.
+ ///
+ /// [jobId] - Identifies the workflow job this worker belongs to.
+ ///
+ /// Completes with a [LeaseWorkItemResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<LeaseWorkItemResponse> lease(
+ LeaseWorkItemRequest request, core.String projectId, core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -747,38 +778,42 @@ class ProjectsJobsWorkItemsResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/workItems:lease';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/workItems:lease';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new LeaseWorkItemResponse.fromJson(data));
}
- /**
- * Reports the status of dataflow WorkItems leased by a worker.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project which owns the WorkItem's job.
- *
- * [jobId] - The job which the WorkItem is part of.
- *
- * Completes with a [ReportWorkItemStatusResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ReportWorkItemStatusResponse> reportStatus(ReportWorkItemStatusRequest request, core.String projectId, core.String jobId) {
+ /// Reports the status of dataflow WorkItems leased by a worker.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project which owns the WorkItem's job.
+ ///
+ /// [jobId] - The job which the WorkItem is part of.
+ ///
+ /// Completes with a [ReportWorkItemStatusResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ReportWorkItemStatusResponse> reportStatus(
+ ReportWorkItemStatusRequest request,
+ core.String projectId,
+ core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -796,50 +831,55 @@ class ProjectsJobsWorkItemsResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/workItems:reportStatus';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/workItems:reportStatus';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new ReportWorkItemStatusResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new ReportWorkItemStatusResponse.fromJson(data));
}
-
}
-
class ProjectsLocationsResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsJobsResourceApi get jobs => new ProjectsLocationsJobsResourceApi(_requester);
- ProjectsLocationsTemplatesResourceApi get templates => new ProjectsLocationsTemplatesResourceApi(_requester);
-
- ProjectsLocationsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Send a worker_message to the service.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project to send the WorkerMessages to.
- *
- * [location] - The location which contains the job
- *
- * Completes with a [SendWorkerMessagesResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<SendWorkerMessagesResponse> workerMessages(SendWorkerMessagesRequest request, core.String projectId, core.String location) {
+ ProjectsLocationsJobsResourceApi get jobs =>
+ new ProjectsLocationsJobsResourceApi(_requester);
+ ProjectsLocationsTemplatesResourceApi get templates =>
+ new ProjectsLocationsTemplatesResourceApi(_requester);
+
+ ProjectsLocationsResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Send a worker_message to the service.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project to send the WorkerMessages to.
+ ///
+ /// [location] - The location which contains the job
+ ///
+ /// Completes with a [SendWorkerMessagesResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<SendWorkerMessagesResponse> workerMessages(
+ SendWorkerMessagesRequest request,
+ core.String projectId,
+ core.String location) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -857,60 +897,66 @@ class ProjectsLocationsResourceApi {
throw new core.ArgumentError("Parameter location is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/WorkerMessages';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/WorkerMessages';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new SendWorkerMessagesResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new SendWorkerMessagesResponse.fromJson(data));
}
-
}
-
class ProjectsLocationsJobsResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsJobsDebugResourceApi get debug => new ProjectsLocationsJobsDebugResourceApi(_requester);
- ProjectsLocationsJobsMessagesResourceApi get messages => new ProjectsLocationsJobsMessagesResourceApi(_requester);
- ProjectsLocationsJobsWorkItemsResourceApi get workItems => new ProjectsLocationsJobsWorkItemsResourceApi(_requester);
-
- ProjectsLocationsJobsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Creates a Cloud Dataflow job.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [location] - The location that contains this job.
- *
- * [replaceJobId] - Deprecated. This field is now in the Job message.
- *
- * [view] - The level of information requested in response.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> create(Job request, core.String projectId, core.String location, {core.String replaceJobId, core.String view}) {
+ ProjectsLocationsJobsDebugResourceApi get debug =>
+ new ProjectsLocationsJobsDebugResourceApi(_requester);
+ ProjectsLocationsJobsMessagesResourceApi get messages =>
+ new ProjectsLocationsJobsMessagesResourceApi(_requester);
+ ProjectsLocationsJobsWorkItemsResourceApi get workItems =>
+ new ProjectsLocationsJobsWorkItemsResourceApi(_requester);
+
+ ProjectsLocationsJobsResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Creates a Cloud Dataflow job.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [replaceJobId] - Deprecated. This field is now in the Job message.
+ ///
+ /// [view] - The level of information requested in response.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> create(
+ Job request, core.String projectId, core.String location,
+ {core.String replaceJobId, core.String view}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -934,45 +980,49 @@ class ProjectsLocationsJobsResourceApi {
_queryParams["view"] = [view];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Gets the state of the specified Cloud Dataflow job.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [location] - The location that contains this job.
- *
- * [jobId] - The job ID.
- *
- * [view] - The level of information requested in response.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> get(core.String projectId, core.String location, core.String jobId, {core.String view}) {
+ /// Gets the state of the specified Cloud Dataflow job.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [jobId] - The job ID.
+ ///
+ /// [view] - The level of information requested in response.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> get(
+ core.String projectId, core.String location, core.String jobId,
+ {core.String view}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -993,41 +1043,45 @@ class ProjectsLocationsJobsResourceApi {
_queryParams["view"] = [view];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId');
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId');
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Request the job status.
- *
- * Request parameters:
- *
- * [projectId] - A project id.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [jobId] - The job to get messages for.
- *
- * [startTime] - Return only metric data that has changed since this time.
- * Default is to return all information about all metrics for the job.
- *
- * Completes with a [JobMetrics].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<JobMetrics> getMetrics(core.String projectId, core.String location, core.String jobId, {core.String startTime}) {
+ /// Request the job status.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - A project id.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [jobId] - The job to get messages for.
+ ///
+ /// [startTime] - Return only metric data that has changed since this time.
+ /// Default is to return all information about all metrics for the job.
+ ///
+ /// Completes with a [JobMetrics].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<JobMetrics> getMetrics(
+ core.String projectId, core.String location, core.String jobId,
+ {core.String startTime}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1048,59 +1102,67 @@ class ProjectsLocationsJobsResourceApi {
_queryParams["startTime"] = [startTime];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/metrics';
-
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/metrics';
+
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new JobMetrics.fromJson(data));
}
- /**
- * List the jobs of a project in a given region.
- *
- * Request parameters:
- *
- * [projectId] - The project which owns the jobs.
- *
- * [location] - The location that contains this job.
- *
- * [filter] - The kind of filter to use.
- * Possible string values are:
- * - "UNKNOWN" : A UNKNOWN.
- * - "ALL" : A ALL.
- * - "TERMINATED" : A TERMINATED.
- * - "ACTIVE" : A ACTIVE.
- *
- * [pageToken] - Set this to the 'next_page_token' field of a previous
- * response
- * to request additional results in a long list.
- *
- * [pageSize] - If there are many jobs, limit response to at most this many.
- * The actual number of jobs returned will be the lesser of max_responses
- * and an unspecified server-defined limit.
- *
- * [view] - Level of information requested in response. Default is
- * `JOB_VIEW_SUMMARY`.
- * Possible string values are:
- * - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
- * - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
- * - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
- * - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
- *
- * Completes with a [ListJobsResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListJobsResponse> list(core.String projectId, core.String location, {core.String filter, core.String pageToken, core.int pageSize, core.String view}) {
+ /// List the jobs of a project in a given region.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project which owns the jobs.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [pageToken] - Set this to the 'next_page_token' field of a previous
+ /// response
+ /// to request additional results in a long list.
+ ///
+ /// [pageSize] - If there are many jobs, limit response to at most this many.
+ /// The actual number of jobs returned will be the lesser of max_responses
+ /// and an unspecified server-defined limit.
+ ///
+ /// [view] - Level of information requested in response. Default is
+ /// `JOB_VIEW_SUMMARY`.
+ /// Possible string values are:
+ /// - "JOB_VIEW_UNKNOWN" : A JOB_VIEW_UNKNOWN.
+ /// - "JOB_VIEW_SUMMARY" : A JOB_VIEW_SUMMARY.
+ /// - "JOB_VIEW_ALL" : A JOB_VIEW_ALL.
+ /// - "JOB_VIEW_DESCRIPTION" : A JOB_VIEW_DESCRIPTION.
+ ///
+ /// [filter] - The kind of filter to use.
+ /// Possible string values are:
+ /// - "UNKNOWN" : A UNKNOWN.
+ /// - "ALL" : A ALL.
+ /// - "TERMINATED" : A TERMINATED.
+ /// - "ACTIVE" : A ACTIVE.
+ ///
+ /// Completes with a [ListJobsResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListJobsResponse> list(
+ core.String projectId, core.String location,
+ {core.String pageToken,
+ core.int pageSize,
+ core.String view,
+ core.String filter}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1114,9 +1176,6 @@ class ProjectsLocationsJobsResourceApi {
if (location == null) {
throw new core.ArgumentError("Parameter location is required.");
}
- if (filter != null) {
- _queryParams["filter"] = [filter];
- }
if (pageToken != null) {
_queryParams["pageToken"] = [pageToken];
}
@@ -1126,41 +1185,47 @@ class ProjectsLocationsJobsResourceApi {
if (view != null) {
_queryParams["view"] = [view];
}
+ if (filter != null) {
+ _queryParams["filter"] = [filter];
+ }
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListJobsResponse.fromJson(data));
}
- /**
- * Updates the state of an existing Cloud Dataflow job.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The ID of the Cloud Platform project that the job belongs to.
- *
- * [location] - The location that contains this job.
- *
- * [jobId] - The job ID.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> update(Job request, core.String projectId, core.String location, core.String jobId) {
+ /// Updates the state of an existing Cloud Dataflow job.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The ID of the Cloud Platform project that the job belongs
+ /// to.
+ ///
+ /// [location] - The location that contains this job.
+ ///
+ /// [jobId] - The job ID.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> update(Job request, core.String projectId,
+ core.String location, core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1181,49 +1246,50 @@ class ProjectsLocationsJobsResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId');
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId');
- var _response = _requester.request(_url,
- "PUT",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
-
}
-
class ProjectsLocationsJobsDebugResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsJobsDebugResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Get encoded debug configuration for component. Not cacheable.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project id.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [jobId] - The job id.
- *
- * Completes with a [GetDebugConfigResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<GetDebugConfigResponse> getConfig(GetDebugConfigRequest request, core.String projectId, core.String location, core.String jobId) {
+ ProjectsLocationsJobsDebugResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Get encoded debug configuration for component. Not cacheable.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project id.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [jobId] - The job id.
+ ///
+ /// Completes with a [GetDebugConfigResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<GetDebugConfigResponse> getConfig(GetDebugConfigRequest request,
+ core.String projectId, core.String location, core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1244,40 +1310,47 @@ class ProjectsLocationsJobsDebugResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/debug/getConfig';
-
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/debug/getConfig';
+
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new GetDebugConfigResponse.fromJson(data));
}
- /**
- * Send encoded debug capture data for component.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project id.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [jobId] - The job id.
- *
- * Completes with a [SendDebugCaptureResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<SendDebugCaptureResponse> sendCapture(SendDebugCaptureRequest request, core.String projectId, core.String location, core.String jobId) {
+ /// Send encoded debug capture data for component.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project id.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [jobId] - The job id.
+ ///
+ /// Completes with a [SendDebugCaptureResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<SendDebugCaptureResponse> sendCapture(
+ SendDebugCaptureRequest request,
+ core.String projectId,
+ core.String location,
+ core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1298,73 +1371,82 @@ class ProjectsLocationsJobsDebugResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/debug/sendCapture';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/debug/sendCapture';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new SendDebugCaptureResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new SendDebugCaptureResponse.fromJson(data));
}
-
}
-
class ProjectsLocationsJobsMessagesResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsJobsMessagesResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Request the job status.
- *
- * Request parameters:
- *
- * [projectId] - A project id.
- *
- * [location] - The location which contains the job specified by job_id.
- *
- * [jobId] - The job to get messages about.
- *
- * [endTime] - Return only messages with timestamps < end_time. The default is
- * now
- * (i.e. return up to the latest messages available).
- *
- * [startTime] - If specified, return only messages with timestamps >=
- * start_time.
- * The default is the job creation time (i.e. beginning of messages).
- *
- * [pageToken] - If supplied, this should be the value of next_page_token
- * returned
- * by an earlier call. This will cause the next page of results to
- * be returned.
- *
- * [pageSize] - If specified, determines the maximum number of messages to
- * return. If unspecified, the service may choose an appropriate
- * default, or may return an arbitrarily large number of results.
- *
- * [minimumImportance] - Filter to only get messages with importance >= level
- * Possible string values are:
- * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN.
- * - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG.
- * - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED.
- * - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC.
- * - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING.
- * - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR.
- *
- * Completes with a [ListJobMessagesResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListJobMessagesResponse> list(core.String projectId, core.String location, core.String jobId, {core.String endTime, core.String startTime, core.String pageToken, core.int pageSize, core.String minimumImportance}) {
+ ProjectsLocationsJobsMessagesResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Request the job status.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - A project id.
+ ///
+ /// [location] - The location which contains the job specified by job_id.
+ ///
+ /// [jobId] - The job to get messages about.
+ ///
+ /// [endTime] - Return only messages with timestamps < end_time. The default
+ /// is now
+ /// (i.e. return up to the latest messages available).
+ ///
+ /// [pageToken] - If supplied, this should be the value of next_page_token
+ /// returned
+ /// by an earlier call. This will cause the next page of results to
+ /// be returned.
+ ///
+ /// [startTime] - If specified, return only messages with timestamps >=
+ /// start_time.
+ /// The default is the job creation time (i.e. beginning of messages).
+ ///
+ /// [pageSize] - If specified, determines the maximum number of messages to
+ /// return. If unspecified, the service may choose an appropriate
+ /// default, or may return an arbitrarily large number of results.
+ ///
+ /// [minimumImportance] - Filter to only get messages with importance >=
+ /// level
+ /// Possible string values are:
+ /// - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : A JOB_MESSAGE_IMPORTANCE_UNKNOWN.
+ /// - "JOB_MESSAGE_DEBUG" : A JOB_MESSAGE_DEBUG.
+ /// - "JOB_MESSAGE_DETAILED" : A JOB_MESSAGE_DETAILED.
+ /// - "JOB_MESSAGE_BASIC" : A JOB_MESSAGE_BASIC.
+ /// - "JOB_MESSAGE_WARNING" : A JOB_MESSAGE_WARNING.
+ /// - "JOB_MESSAGE_ERROR" : A JOB_MESSAGE_ERROR.
+ ///
+ /// Completes with a [ListJobMessagesResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListJobMessagesResponse> list(
+ core.String projectId, core.String location, core.String jobId,
+ {core.String endTime,
+ core.String pageToken,
+ core.String startTime,
+ core.int pageSize,
+ core.String minimumImportance}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1384,12 +1466,12 @@ class ProjectsLocationsJobsMessagesResourceApi {
if (endTime != null) {
_queryParams["endTime"] = [endTime];
}
- if (startTime != null) {
- _queryParams["startTime"] = [startTime];
- }
if (pageToken != null) {
_queryParams["pageToken"] = [pageToken];
}
+ if (startTime != null) {
+ _queryParams["startTime"] = [startTime];
+ }
if (pageSize != null) {
_queryParams["pageSize"] = ["${pageSize}"];
}
@@ -1397,49 +1479,51 @@ class ProjectsLocationsJobsMessagesResourceApi {
_queryParams["minimumImportance"] = [minimumImportance];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/messages';
-
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/messages';
+
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListJobMessagesResponse.fromJson(data));
}
-
}
-
class ProjectsLocationsJobsWorkItemsResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsJobsWorkItemsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Leases a dataflow WorkItem to run.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Identifies the project this worker belongs to.
- *
- * [location] - The location which contains the WorkItem's job.
- *
- * [jobId] - Identifies the workflow job this worker belongs to.
- *
- * Completes with a [LeaseWorkItemResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<LeaseWorkItemResponse> lease(LeaseWorkItemRequest request, core.String projectId, core.String location, core.String jobId) {
+ ProjectsLocationsJobsWorkItemsResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Leases a dataflow WorkItem to run.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Identifies the project this worker belongs to.
+ ///
+ /// [location] - The location which contains the WorkItem's job.
+ ///
+ /// [jobId] - Identifies the workflow job this worker belongs to.
+ ///
+ /// Completes with a [LeaseWorkItemResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<LeaseWorkItemResponse> lease(LeaseWorkItemRequest request,
+ core.String projectId, core.String location, core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1460,40 +1544,47 @@ class ProjectsLocationsJobsWorkItemsResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/workItems:lease';
-
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/workItems:lease';
+
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new LeaseWorkItemResponse.fromJson(data));
}
- /**
- * Reports the status of dataflow WorkItems leased by a worker.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - The project which owns the WorkItem's job.
- *
- * [location] - The location which contains the WorkItem's job.
- *
- * [jobId] - The job which the WorkItem is part of.
- *
- * Completes with a [ReportWorkItemStatusResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ReportWorkItemStatusResponse> reportStatus(ReportWorkItemStatusRequest request, core.String projectId, core.String location, core.String jobId) {
+ /// Reports the status of dataflow WorkItems leased by a worker.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - The project which owns the WorkItem's job.
+ ///
+ /// [location] - The location which contains the WorkItem's job.
+ ///
+ /// [jobId] - The job which the WorkItem is part of.
+ ///
+ /// Completes with a [ReportWorkItemStatusResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ReportWorkItemStatusResponse> reportStatus(
+ ReportWorkItemStatusRequest request,
+ core.String projectId,
+ core.String location,
+ core.String jobId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1514,48 +1605,51 @@ class ProjectsLocationsJobsWorkItemsResourceApi {
throw new core.ArgumentError("Parameter jobId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/jobs/' + commons.Escaper.ecapeVariable('$jobId') + '/workItems:reportStatus';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/jobs/' +
+ commons.Escaper.ecapeVariable('$jobId') +
+ '/workItems:reportStatus';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new ReportWorkItemStatusResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new ReportWorkItemStatusResponse.fromJson(data));
}
-
}
-
class ProjectsLocationsTemplatesResourceApi {
final commons.ApiRequester _requester;
- ProjectsLocationsTemplatesResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Creates a Cloud Dataflow job from a template.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * [location] - The location to which to direct the request.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> create(CreateJobFromTemplateRequest request, core.String projectId, core.String location) {
+ ProjectsLocationsTemplatesResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Creates a Cloud Dataflow job from a template.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// [location] - The location to which to direct the request.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> create(CreateJobFromTemplateRequest request,
+ core.String projectId, core.String location) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1573,45 +1667,48 @@ class ProjectsLocationsTemplatesResourceApi {
throw new core.ArgumentError("Parameter location is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/templates';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/templates';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Get the template associated with a template.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * [location] - The location to which to direct the request.
- *
- * [view] - The view to retrieve. Defaults to METADATA_ONLY.
- * Possible string values are:
- * - "METADATA_ONLY" : A METADATA_ONLY.
- *
- * [gcsPath] - Required. A Cloud Storage path to the template from which to
- * create the job.
- * Must be a valid Cloud Storage URL, beginning with `gs://`.
- *
- * Completes with a [GetTemplateResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<GetTemplateResponse> get(core.String projectId, core.String location, {core.String view, core.String gcsPath}) {
+ /// Get the template associated with a template.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// [location] - The location to which to direct the request.
+ ///
+ /// [view] - The view to retrieve. Defaults to METADATA_ONLY.
+ /// Possible string values are:
+ /// - "METADATA_ONLY" : A METADATA_ONLY.
+ ///
+ /// [gcsPath] - Required. A Cloud Storage path to the template from which to
+ /// create the job.
+ /// Must be a valid Cloud Storage URL, beginning with `gs://`.
+ ///
+ /// Completes with a [GetTemplateResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<GetTemplateResponse> get(
+ core.String projectId, core.String location,
+ {core.String view, core.String gcsPath}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1632,48 +1729,51 @@ class ProjectsLocationsTemplatesResourceApi {
_queryParams["gcsPath"] = [gcsPath];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/templates:get';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/templates:get';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new GetTemplateResponse.fromJson(data));
}
- /**
- * Launch a template.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * [location] - The location to which to direct the request.
- *
- * [gcsPath] - Required. A Cloud Storage path to the template from which to
- * create
- * the job.
- * Must be valid Cloud Storage URL, beginning with 'gs://'.
- *
- * [validateOnly] - If true, the request is validated but not actually
- * executed.
- * Defaults to false.
- *
- * Completes with a [LaunchTemplateResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<LaunchTemplateResponse> launch(LaunchTemplateParameters request, core.String projectId, core.String location, {core.String gcsPath, core.bool validateOnly}) {
+ /// Launch a template.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// [location] - The location to which to direct the request.
+ ///
+ /// [validateOnly] - If true, the request is validated but not actually
+ /// executed.
+ /// Defaults to false.
+ ///
+ /// [gcsPath] - Required. A Cloud Storage path to the template from which to
+ /// create
+ /// the job.
+ /// Must be valid Cloud Storage URL, beginning with 'gs://'.
+ ///
+ /// Completes with a [LaunchTemplateResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<LaunchTemplateResponse> launch(LaunchTemplateParameters request,
+ core.String projectId, core.String location,
+ {core.bool validateOnly, core.String gcsPath}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1690,53 +1790,53 @@ class ProjectsLocationsTemplatesResourceApi {
if (location == null) {
throw new core.ArgumentError("Parameter location is required.");
}
- if (gcsPath != null) {
- _queryParams["gcsPath"] = [gcsPath];
- }
if (validateOnly != null) {
_queryParams["validateOnly"] = ["${validateOnly}"];
}
+ if (gcsPath != null) {
+ _queryParams["gcsPath"] = [gcsPath];
+ }
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/locations/' + commons.Escaper.ecapeVariable('$location') + '/templates:launch';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/locations/' +
+ commons.Escaper.ecapeVariable('$location') +
+ '/templates:launch';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new LaunchTemplateResponse.fromJson(data));
}
-
}
-
class ProjectsTemplatesResourceApi {
final commons.ApiRequester _requester;
- ProjectsTemplatesResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Creates a Cloud Dataflow job from a template.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * Completes with a [Job].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Job> create(CreateJobFromTemplateRequest request, core.String projectId) {
+ ProjectsTemplatesResourceApi(commons.ApiRequester client)
+ : _requester = client;
+
+ /// Creates a Cloud Dataflow job from a template.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// Completes with a [Job].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Job> create(
+ CreateJobFromTemplateRequest request, core.String projectId) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1751,45 +1851,45 @@ class ProjectsTemplatesResourceApi {
throw new core.ArgumentError("Parameter projectId is required.");
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/templates';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/templates';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Job.fromJson(data));
}
- /**
- * Get the template associated with a template.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * [location] - The location to which to direct the request.
- *
- * [view] - The view to retrieve. Defaults to METADATA_ONLY.
- * Possible string values are:
- * - "METADATA_ONLY" : A METADATA_ONLY.
- *
- * [gcsPath] - Required. A Cloud Storage path to the template from which to
- * create the job.
- * Must be a valid Cloud Storage URL, beginning with `gs://`.
- *
- * Completes with a [GetTemplateResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<GetTemplateResponse> get(core.String projectId, {core.String location, core.String view, core.String gcsPath}) {
+ /// Get the template associated with a template.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// [view] - The view to retrieve. Defaults to METADATA_ONLY.
+ /// Possible string values are:
+ /// - "METADATA_ONLY" : A METADATA_ONLY.
+ ///
+ /// [gcsPath] - Required. A Cloud Storage path to the template from which to
+ /// create the job.
+ /// Must be a valid Cloud Storage URL, beginning with `gs://`.
+ ///
+ /// [location] - The location to which to direct the request.
+ ///
+ /// Completes with a [GetTemplateResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<GetTemplateResponse> get(core.String projectId,
+ {core.String view, core.String gcsPath, core.String location}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1800,58 +1900,59 @@ class ProjectsTemplatesResourceApi {
if (projectId == null) {
throw new core.ArgumentError("Parameter projectId is required.");
}
- if (location != null) {
- _queryParams["location"] = [location];
- }
if (view != null) {
_queryParams["view"] = [view];
}
if (gcsPath != null) {
_queryParams["gcsPath"] = [gcsPath];
}
+ if (location != null) {
+ _queryParams["location"] = [location];
+ }
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/templates:get';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/templates:get';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new GetTemplateResponse.fromJson(data));
}
- /**
- * Launch a template.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [projectId] - Required. The ID of the Cloud Platform project that the job
- * belongs to.
- *
- * [location] - The location to which to direct the request.
- *
- * [validateOnly] - If true, the request is validated but not actually
- * executed.
- * Defaults to false.
- *
- * [gcsPath] - Required. A Cloud Storage path to the template from which to
- * create
- * the job.
- * Must be valid Cloud Storage URL, beginning with 'gs://'.
- *
- * Completes with a [LaunchTemplateResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<LaunchTemplateResponse> launch(LaunchTemplateParameters request, core.String projectId, {core.String location, core.bool validateOnly, core.String gcsPath}) {
+ /// Launch a template.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [projectId] - Required. The ID of the Cloud Platform project that the job
+ /// belongs to.
+ ///
+ /// [location] - The location to which to direct the request.
+ ///
+ /// [validateOnly] - If true, the request is validated but not actually
+ /// executed.
+ /// Defaults to false.
+ ///
+ /// [gcsPath] - Required. A Cloud Storage path to the template from which to
+ /// create
+ /// the job.
+ /// Must be valid Cloud Storage URL, beginning with 'gs://'.
+ ///
+ /// Completes with a [LaunchTemplateResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<LaunchTemplateResponse> launch(
+ LaunchTemplateParameters request, core.String projectId,
+ {core.String location, core.bool validateOnly, core.String gcsPath}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -1875,31 +1976,30 @@ class ProjectsTemplatesResourceApi {
_queryParams["gcsPath"] = [gcsPath];
}
- _url = 'v1b3/projects/' + commons.Escaper.ecapeVariable('$projectId') + '/templates:launch';
+ _url = 'v1b3/projects/' +
+ commons.Escaper.ecapeVariable('$projectId') +
+ '/templates:launch';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new LaunchTemplateResponse.fromJson(data));
}
-
}
-
-
-/**
- * Obsolete in favor of ApproximateReportedProgress and ApproximateSplitRequest.
- */
+/// Obsolete in favor of ApproximateReportedProgress and
+/// ApproximateSplitRequest.
class ApproximateProgress {
- /** Obsolete. */
+ /// Obsolete.
core.double percentComplete;
- /** Obsolete. */
+
+ /// Obsolete.
Position position;
- /** Obsolete. */
+
+ /// Obsolete.
core.String remainingTime;
ApproximateProgress();
@@ -1917,7 +2017,8 @@ class ApproximateProgress {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (percentComplete != null) {
_json["percentComplete"] = percentComplete;
}
@@ -1931,58 +2032,58 @@ class ApproximateProgress {
}
}
-/** A progress measurement of a WorkItem by a worker. */
+/// A progress measurement of a WorkItem by a worker.
class ApproximateReportedProgress {
- /**
- * Total amount of parallelism in the portion of input of this task that has
- * already been consumed and is no longer active. In the first two examples
- * above (see remaining_parallelism), the value should be 29 or 2
- * respectively. The sum of remaining_parallelism and consumed_parallelism
- * should equal the total amount of parallelism in this work item. If
- * specified, must be finite.
- */
+ /// Total amount of parallelism in the portion of input of this task that has
+ /// already been consumed and is no longer active. In the first two examples
+ /// above (see remaining_parallelism), the value should be 29 or 2
+ /// respectively. The sum of remaining_parallelism and consumed_parallelism
+ /// should equal the total amount of parallelism in this work item. If
+ /// specified, must be finite.
ReportedParallelism consumedParallelism;
- /**
- * Completion as fraction of the input consumed, from 0.0 (beginning, nothing
- * consumed), to 1.0 (end of the input, entire input consumed).
- */
+
+ /// Completion as fraction of the input consumed, from 0.0 (beginning,
+ /// nothing
+ /// consumed), to 1.0 (end of the input, entire input consumed).
core.double fractionConsumed;
- /** A Position within the work to represent a progress. */
+
+ /// A Position within the work to represent a progress.
Position position;
- /**
- * Total amount of parallelism in the input of this task that remains,
- * (i.e. can be delegated to this task and any new tasks via dynamic
- * splitting). Always at least 1 for non-finished work items and 0 for
- * finished.
- *
- * "Amount of parallelism" refers to how many non-empty parts of the input
- * can be read in parallel. This does not necessarily equal number
- * of records. An input that can be read in parallel down to the
- * individual records is called "perfectly splittable".
- * An example of non-perfectly parallelizable input is a block-compressed
- * file format where a block of records has to be read as a whole,
- * but different blocks can be read in parallel.
- *
- * Examples:
- * * If we are processing record #30 (starting at 1) out of 50 in a perfectly
- * splittable 50-record input, this value should be 21 (20 remaining + 1
- * current).
- * * If we are reading through block 3 in a block-compressed file consisting
- * of 5 blocks, this value should be 3 (since blocks 4 and 5 can be
- * processed in parallel by new tasks via dynamic splitting and the current
- * task remains processing block 3).
- * * If we are reading through the last block in a block-compressed file,
- * or reading or processing the last record in a perfectly splittable
- * input, this value should be 1, because apart from the current task, no
- * additional remainder can be split off.
- */
+
+ /// Total amount of parallelism in the input of this task that remains,
+ /// (i.e. can be delegated to this task and any new tasks via dynamic
+ /// splitting). Always at least 1 for non-finished work items and 0 for
+ /// finished.
+ ///
+ /// "Amount of parallelism" refers to how many non-empty parts of the input
+ /// can be read in parallel. This does not necessarily equal number
+ /// of records. An input that can be read in parallel down to the
+ /// individual records is called "perfectly splittable".
+ /// An example of non-perfectly parallelizable input is a block-compressed
+ /// file format where a block of records has to be read as a whole,
+ /// but different blocks can be read in parallel.
+ ///
+ /// Examples:
+ /// * If we are processing record #30 (starting at 1) out of 50 in a
+ /// perfectly
+ /// splittable 50-record input, this value should be 21 (20 remaining + 1
+ /// current).
+ /// * If we are reading through block 3 in a block-compressed file consisting
+ /// of 5 blocks, this value should be 3 (since blocks 4 and 5 can be
+ /// processed in parallel by new tasks via dynamic splitting and the current
+ /// task remains processing block 3).
+ /// * If we are reading through the last block in a block-compressed file,
+ /// or reading or processing the last record in a perfectly splittable
+ /// input, this value should be 1, because apart from the current task, no
+ /// additional remainder can be split off.
ReportedParallelism remainingParallelism;
ApproximateReportedProgress();
ApproximateReportedProgress.fromJson(core.Map _json) {
if (_json.containsKey("consumedParallelism")) {
- consumedParallelism = new ReportedParallelism.fromJson(_json["consumedParallelism"]);
+ consumedParallelism =
+ new ReportedParallelism.fromJson(_json["consumedParallelism"]);
}
if (_json.containsKey("fractionConsumed")) {
fractionConsumed = _json["fractionConsumed"];
@@ -1991,12 +2092,14 @@ class ApproximateReportedProgress {
position = new Position.fromJson(_json["position"]);
}
if (_json.containsKey("remainingParallelism")) {
- remainingParallelism = new ReportedParallelism.fromJson(_json["remainingParallelism"]);
+ remainingParallelism =
+ new ReportedParallelism.fromJson(_json["remainingParallelism"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (consumedParallelism != null) {
_json["consumedParallelism"] = (consumedParallelism).toJson();
}
@@ -2013,16 +2116,14 @@ class ApproximateReportedProgress {
}
}
-/**
- * A suggestion by the service to the worker to dynamically split the WorkItem.
- */
+/// A suggestion by the service to the worker to dynamically split the
+/// WorkItem.
class ApproximateSplitRequest {
- /**
- * A fraction at which to split the work item, from 0.0 (beginning of the
- * input) to 1.0 (end of the input).
- */
+ /// A fraction at which to split the work item, from 0.0 (beginning of the
+ /// input) to 1.0 (end of the input).
core.double fractionConsumed;
- /** A Position at which to split the work item. */
+
+ /// A Position at which to split the work item.
Position position;
ApproximateSplitRequest();
@@ -2037,7 +2138,8 @@ class ApproximateSplitRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (fractionConsumed != null) {
_json["fractionConsumed"] = fractionConsumed;
}
@@ -2048,49 +2150,46 @@ class ApproximateSplitRequest {
}
}
-/**
- * A structured message reporting an autoscaling decision made by the Dataflow
- * service.
- */
+/// A structured message reporting an autoscaling decision made by the Dataflow
+/// service.
class AutoscalingEvent {
- /** The current number of workers the job has. */
+ /// The current number of workers the job has.
core.String currentNumWorkers;
- /**
- * A message describing why the system decided to adjust the current
- * number of workers, why it failed, or why the system decided to
- * not make any changes to the number of workers.
- */
+
+ /// A message describing why the system decided to adjust the current
+ /// number of workers, why it failed, or why the system decided to
+ /// not make any changes to the number of workers.
StructuredMessage description;
- /**
- * The type of autoscaling event to report.
- * Possible string values are:
- * - "TYPE_UNKNOWN" : Default type for the enum. Value should never be
- * returned.
- * - "TARGET_NUM_WORKERS_CHANGED" : The TARGET_NUM_WORKERS_CHANGED type should
- * be used when the target
- * worker pool size has changed at the start of an actuation. An event
- * should always be specified as TARGET_NUM_WORKERS_CHANGED if it reflects
- * a change in the target_num_workers.
- * - "CURRENT_NUM_WORKERS_CHANGED" : The CURRENT_NUM_WORKERS_CHANGED type
- * should be used when actual worker
- * pool size has been changed, but the target_num_workers has not changed.
- * - "ACTUATION_FAILURE" : The ACTUATION_FAILURE type should be used when we
- * want to report
- * an error to the user indicating why the current number of workers
- * in the pool could not be changed.
- * Displayed in the current status and history widgets.
- * - "NO_CHANGE" : Used when we want to report to the user a reason why we are
- * not currently adjusting the number of workers.
- * Should specify both target_num_workers, current_num_workers and a
- * decision_message.
- */
+
+ /// The type of autoscaling event to report.
+ /// Possible string values are:
+ /// - "TYPE_UNKNOWN" : Default type for the enum. Value should never be
+ /// returned.
+ /// - "TARGET_NUM_WORKERS_CHANGED" : The TARGET_NUM_WORKERS_CHANGED type
+ /// should be used when the target
+ /// worker pool size has changed at the start of an actuation. An event
+ /// should always be specified as TARGET_NUM_WORKERS_CHANGED if it reflects
+ /// a change in the target_num_workers.
+ /// - "CURRENT_NUM_WORKERS_CHANGED" : The CURRENT_NUM_WORKERS_CHANGED type
+ /// should be used when actual worker
+ /// pool size has been changed, but the target_num_workers has not changed.
+ /// - "ACTUATION_FAILURE" : The ACTUATION_FAILURE type should be used when we
+ /// want to report
+ /// an error to the user indicating why the current number of workers
+ /// in the pool could not be changed.
+ /// Displayed in the current status and history widgets.
+ /// - "NO_CHANGE" : Used when we want to report to the user a reason why we
+ /// are
+ /// not currently adjusting the number of workers.
+ /// Should specify both target_num_workers, current_num_workers and a
+ /// decision_message.
core.String eventType;
- /** The target number of workers the worker pool wants to resize to use. */
+
+ /// The target number of workers the worker pool wants to resize to use.
core.String targetNumWorkers;
- /**
- * The time this event was emitted to indicate a new target or current
- * num_workers value.
- */
+
+ /// The time this event was emitted to indicate a new target or current
+ /// num_workers value.
core.String time;
AutoscalingEvent();
@@ -2114,7 +2213,8 @@ class AutoscalingEvent {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (currentNumWorkers != null) {
_json["currentNumWorkers"] = currentNumWorkers;
}
@@ -2134,19 +2234,18 @@ class AutoscalingEvent {
}
}
-/** Settings for WorkerPool autoscaling. */
+/// Settings for WorkerPool autoscaling.
class AutoscalingSettings {
- /**
- * The algorithm to use for autoscaling.
- * Possible string values are:
- * - "AUTOSCALING_ALGORITHM_UNKNOWN" : The algorithm is unknown, or
- * unspecified.
- * - "AUTOSCALING_ALGORITHM_NONE" : Disable autoscaling.
- * - "AUTOSCALING_ALGORITHM_BASIC" : Increase worker count over time to reduce
- * job execution time.
- */
+ /// The algorithm to use for autoscaling.
+ /// Possible string values are:
+ /// - "AUTOSCALING_ALGORITHM_UNKNOWN" : The algorithm is unknown, or
+ /// unspecified.
+ /// - "AUTOSCALING_ALGORITHM_NONE" : Disable autoscaling.
+ /// - "AUTOSCALING_ALGORITHM_BASIC" : Increase worker count over time to
+ /// reduce job execution time.
core.String algorithm;
- /** The maximum number of workers to cap scaling at. */
+
+ /// The maximum number of workers to cap scaling at.
core.int maxNumWorkers;
AutoscalingSettings();
@@ -2161,7 +2260,8 @@ class AutoscalingSettings {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (algorithm != null) {
_json["algorithm"] = algorithm;
}
@@ -2172,19 +2272,17 @@ class AutoscalingSettings {
}
}
-/** Modeled after information exposed by /proc/stat. */
+/// Modeled after information exposed by /proc/stat.
class CPUTime {
- /**
- * Average CPU utilization rate (% non-idle cpu / second) since previous
- * sample.
- */
+ /// Average CPU utilization rate (% non-idle cpu / second) since previous
+ /// sample.
core.double rate;
- /** Timestamp of the measurement. */
+
+ /// Timestamp of the measurement.
core.String timestamp;
- /**
- * Total active CPU time across all cores (ie., non-idle) in milliseconds
- * since start-up.
- */
+
+ /// Total active CPU time across all cores (ie., non-idle) in milliseconds
+ /// since start-up.
core.String totalMs;
CPUTime();
@@ -2202,7 +2300,8 @@ class CPUTime {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (rate != null) {
_json["rate"] = rate;
}
@@ -2216,21 +2315,17 @@ class CPUTime {
}
}
-/**
- * Description of an interstitial value between transforms in an execution
- * stage.
- */
+/// Description of an interstitial value between transforms in an execution
+/// stage.
class ComponentSource {
- /** Dataflow service generated name for this source. */
+ /// Dataflow service generated name for this source.
core.String name;
- /**
- * User name for the original user transform or collection with which this
- * source is most closely associated.
- */
+
+ /// User name for the original user transform or collection with which this
+ /// source is most closely associated.
core.String originalTransformOrCollection;
- /**
- * Human-readable name for this transform; may be user or system generated.
- */
+
+ /// Human-readable name for this transform; may be user or system generated.
core.String userName;
ComponentSource();
@@ -2248,7 +2343,8 @@ class ComponentSource {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (name != null) {
_json["name"] = name;
}
@@ -2262,18 +2358,16 @@ class ComponentSource {
}
}
-/** Description of a transform executed as part of an execution stage. */
+/// Description of a transform executed as part of an execution stage.
class ComponentTransform {
- /** Dataflow service generated name for this source. */
+ /// Dataflow service generated name for this source.
core.String name;
- /**
- * User name for the original user transform with which this transform is
- * most closely associated.
- */
+
+ /// User name for the original user transform with which this transform is
+ /// most closely associated.
core.String originalTransform;
- /**
- * Human-readable name for this transform; may be user or system generated.
- */
+
+ /// Human-readable name for this transform; may be user or system generated.
core.String userName;
ComponentTransform();
@@ -2291,7 +2385,8 @@ class ComponentTransform {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (name != null) {
_json["name"] = name;
}
@@ -2305,19 +2400,24 @@ class ComponentTransform {
}
}
-/** All configuration data for a particular Computation. */
+/// All configuration data for a particular Computation.
class ComputationTopology {
- /** The ID of the computation. */
+ /// The ID of the computation.
core.String computationId;
- /** The inputs to the computation. */
+
+ /// The inputs to the computation.
core.List<StreamLocation> inputs;
- /** The key ranges processed by the computation. */
+
+ /// The key ranges processed by the computation.
core.List<KeyRangeLocation> keyRanges;
- /** The outputs from the computation. */
+
+ /// The outputs from the computation.
core.List<StreamLocation> outputs;
- /** The state family values. */
+
+ /// The state family values.
core.List<StateFamilyConfig> stateFamilies;
- /** The system stage name. */
+
+ /// The system stage name.
core.String systemStageName;
ComputationTopology();
@@ -2327,16 +2427,24 @@ class ComputationTopology {
computationId = _json["computationId"];
}
if (_json.containsKey("inputs")) {
- inputs = _json["inputs"].map((value) => new StreamLocation.fromJson(value)).toList();
+ inputs = _json["inputs"]
+ .map((value) => new StreamLocation.fromJson(value))
+ .toList();
}
if (_json.containsKey("keyRanges")) {
- keyRanges = _json["keyRanges"].map((value) => new KeyRangeLocation.fromJson(value)).toList();
+ keyRanges = _json["keyRanges"]
+ .map((value) => new KeyRangeLocation.fromJson(value))
+ .toList();
}
if (_json.containsKey("outputs")) {
- outputs = _json["outputs"].map((value) => new StreamLocation.fromJson(value)).toList();
+ outputs = _json["outputs"]
+ .map((value) => new StreamLocation.fromJson(value))
+ .toList();
}
if (_json.containsKey("stateFamilies")) {
- stateFamilies = _json["stateFamilies"].map((value) => new StateFamilyConfig.fromJson(value)).toList();
+ stateFamilies = _json["stateFamilies"]
+ .map((value) => new StateFamilyConfig.fromJson(value))
+ .toList();
}
if (_json.containsKey("systemStageName")) {
systemStageName = _json["systemStageName"];
@@ -2344,7 +2452,8 @@ class ComputationTopology {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (computationId != null) {
_json["computationId"] = computationId;
}
@@ -2358,7 +2467,8 @@ class ComputationTopology {
_json["outputs"] = outputs.map((value) => (value).toJson()).toList();
}
if (stateFamilies != null) {
- _json["stateFamilies"] = stateFamilies.map((value) => (value).toJson()).toList();
+ _json["stateFamilies"] =
+ stateFamilies.map((value) => (value).toJson()).toList();
}
if (systemStageName != null) {
_json["systemStageName"] = systemStageName;
@@ -2367,15 +2477,14 @@ class ComputationTopology {
}
}
-/**
- * A position that encapsulates an inner position and an index for the inner
- * position. A ConcatPosition can be used by a reader of a source that
- * encapsulates a set of other sources.
- */
+/// A position that encapsulates an inner position and an index for the inner
+/// position. A ConcatPosition can be used by a reader of a source that
+/// encapsulates a set of other sources.
class ConcatPosition {
- /** Index of the inner source. */
+ /// Index of the inner source.
core.int index;
- /** Position within the inner source. */
+
+ /// Position within the inner source.
Position position;
ConcatPosition();
@@ -2390,7 +2499,8 @@ class ConcatPosition {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (index != null) {
_json["index"] = index;
}
@@ -2401,43 +2511,40 @@ class ConcatPosition {
}
}
-/**
- * CounterMetadata includes all static non-name non-value counter attributes.
- */
+/// CounterMetadata includes all static non-name non-value counter attributes.
class CounterMetadata {
- /** Human-readable description of the counter semantics. */
+ /// Human-readable description of the counter semantics.
core.String description;
- /**
- * Counter aggregation kind.
- * Possible string values are:
- * - "INVALID" : Counter aggregation kind was not set.
- * - "SUM" : Aggregated value is the sum of all contributed values.
- * - "MAX" : Aggregated value is the max of all contributed values.
- * - "MIN" : Aggregated value is the min of all contributed values.
- * - "MEAN" : Aggregated value is the mean of all contributed values.
- * - "OR" : Aggregated value represents the logical 'or' of all contributed
- * values.
- * - "AND" : Aggregated value represents the logical 'and' of all contributed
- * values.
- * - "SET" : Aggregated value is a set of unique contributed values.
- * - "DISTRIBUTION" : Aggregated value captures statistics about a
- * distribution.
- */
+
+ /// Counter aggregation kind.
+ /// Possible string values are:
+ /// - "INVALID" : Counter aggregation kind was not set.
+ /// - "SUM" : Aggregated value is the sum of all contributed values.
+ /// - "MAX" : Aggregated value is the max of all contributed values.
+ /// - "MIN" : Aggregated value is the min of all contributed values.
+ /// - "MEAN" : Aggregated value is the mean of all contributed values.
+ /// - "OR" : Aggregated value represents the logical 'or' of all contributed
+ /// values.
+ /// - "AND" : Aggregated value represents the logical 'and' of all
+ /// contributed values.
+ /// - "SET" : Aggregated value is a set of unique contributed values.
+ /// - "DISTRIBUTION" : Aggregated value captures statistics about a
+ /// distribution.
core.String kind;
- /** A string referring to the unit type. */
+
+ /// A string referring to the unit type.
core.String otherUnits;
- /**
- * System defined Units, see above enum.
- * Possible string values are:
- * - "BYTES" : Counter returns a value in bytes.
- * - "BYTES_PER_SEC" : Counter returns a value in bytes per second.
- * - "MILLISECONDS" : Counter returns a value in milliseconds.
- * - "MICROSECONDS" : Counter returns a value in microseconds.
- * - "NANOSECONDS" : Counter returns a value in nanoseconds.
- * - "TIMESTAMP_MSEC" : Counter returns a timestamp in milliseconds.
- * - "TIMESTAMP_USEC" : Counter returns a timestamp in microseconds.
- * - "TIMESTAMP_NSEC" : Counter returns a timestamp in nanoseconds.
- */
+
+ /// System defined Units, see above enum.
+ /// Possible string values are:
+ /// - "BYTES" : Counter returns a value in bytes.
+ /// - "BYTES_PER_SEC" : Counter returns a value in bytes per second.
+ /// - "MILLISECONDS" : Counter returns a value in milliseconds.
+ /// - "MICROSECONDS" : Counter returns a value in microseconds.
+ /// - "NANOSECONDS" : Counter returns a value in nanoseconds.
+ /// - "TIMESTAMP_MSEC" : Counter returns a timestamp in milliseconds.
+ /// - "TIMESTAMP_USEC" : Counter returns a timestamp in microseconds.
+ /// - "TIMESTAMP_NSEC" : Counter returns a timestamp in nanoseconds.
core.String standardUnits;
CounterMetadata();
@@ -2458,7 +2565,8 @@ class CounterMetadata {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (description != null) {
_json["description"] = description;
}
@@ -2475,46 +2583,50 @@ class CounterMetadata {
}
}
-/**
- * Identifies a counter within a per-job namespace. Counters whose structured
- * names are the same get merged into a single value for the job.
- */
+/// Identifies a counter within a per-job namespace. Counters whose structured
+/// names are the same get merged into a single value for the job.
class CounterStructuredName {
- /** Name of the optimized step being executed by the workers. */
+ /// Name of the optimized step being executed by the workers.
core.String componentStepName;
- /**
- * Name of the stage. An execution step contains multiple component steps.
- */
+
+ /// Name of the stage. An execution step contains multiple component steps.
core.String executionStepName;
- /**
- * Counter name. Not necessarily globally-unique, but unique within the
- * context of the other fields.
- * Required.
- */
+
+ /// Counter name. Not necessarily globally-unique, but unique within the
+ /// context of the other fields.
+ /// Required.
core.String name;
- /**
- * One of the standard Origins defined above.
- * Possible string values are:
- * - "SYSTEM" : Counter was created by the Dataflow system.
- * - "USER" : Counter was created by the user.
- */
+
+ /// One of the standard Origins defined above.
+ /// Possible string values are:
+ /// - "SYSTEM" : Counter was created by the Dataflow system.
+ /// - "USER" : Counter was created by the user.
core.String origin;
- /** A string containing a more specific namespace of the counter's origin. */
+
+ /// A string containing a more specific namespace of the counter's origin.
core.String originNamespace;
- /**
- * System generated name of the original step in the user's graph, before
- * optimization.
- */
+
+ /// The GroupByKey step name from the original graph.
+ core.String originalShuffleStepName;
+
+ /// System generated name of the original step in the user's graph, before
+ /// optimization.
core.String originalStepName;
- /**
- * Portion of this counter, either key or value.
- * Possible string values are:
- * - "ALL" : Counter portion has not been set.
- * - "KEY" : Counter reports a key.
- * - "VALUE" : Counter reports a value.
- */
+
+ /// Portion of this counter, either key or value.
+ /// Possible string values are:
+ /// - "ALL" : Counter portion has not been set.
+ /// - "KEY" : Counter reports a key.
+ /// - "VALUE" : Counter reports a value.
core.String portion;
- /** ID of a particular worker. */
+
+ /// ID of a side input being read from/written to. Side inputs are identified
+ /// by a pair of (reader, input_index). The reader is usually equal to the
+ /// original name, but it may be different, if a ParDo emits it's Iterator /
+ /// Map side input object.
+ SideInputId sideInput;
+
+ /// ID of a particular worker.
core.String workerId;
CounterStructuredName();
@@ -2535,19 +2647,26 @@ class CounterStructuredName {
if (_json.containsKey("originNamespace")) {
originNamespace = _json["originNamespace"];
}
+ if (_json.containsKey("originalShuffleStepName")) {
+ originalShuffleStepName = _json["originalShuffleStepName"];
+ }
if (_json.containsKey("originalStepName")) {
originalStepName = _json["originalStepName"];
}
if (_json.containsKey("portion")) {
portion = _json["portion"];
}
+ if (_json.containsKey("sideInput")) {
+ sideInput = new SideInputId.fromJson(_json["sideInput"]);
+ }
if (_json.containsKey("workerId")) {
workerId = _json["workerId"];
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (componentStepName != null) {
_json["componentStepName"] = componentStepName;
}
@@ -2563,12 +2682,18 @@ class CounterStructuredName {
if (originNamespace != null) {
_json["originNamespace"] = originNamespace;
}
+ if (originalShuffleStepName != null) {
+ _json["originalShuffleStepName"] = originalShuffleStepName;
+ }
if (originalStepName != null) {
_json["originalStepName"] = originalStepName;
}
if (portion != null) {
_json["portion"] = portion;
}
+ if (sideInput != null) {
+ _json["sideInput"] = (sideInput).toJson();
+ }
if (workerId != null) {
_json["workerId"] = workerId;
}
@@ -2576,14 +2701,14 @@ class CounterStructuredName {
}
}
-/**
- * A single message which encapsulates structured name and metadata for a given
- * counter.
- */
+/// A single message which encapsulates structured name and metadata for a
+/// given
+/// counter.
class CounterStructuredNameAndMetadata {
- /** Metadata associated with a counter */
+ /// Metadata associated with a counter
CounterMetadata metadata;
- /** Structured name of the counter. */
+
+ /// Structured name of the counter.
CounterStructuredName name;
CounterStructuredNameAndMetadata();
@@ -2598,7 +2723,8 @@ class CounterStructuredNameAndMetadata {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (metadata != null) {
_json["metadata"] = (metadata).toJson();
}
@@ -2609,49 +2735,56 @@ class CounterStructuredNameAndMetadata {
}
}
-/** An update to a Counter sent from a worker. */
+/// An update to a Counter sent from a worker.
class CounterUpdate {
- /** Boolean value for And, Or. */
+ /// Boolean value for And, Or.
core.bool boolean;
- /**
- * True if this counter is reported as the total cumulative aggregate
- * value accumulated since the worker started working on this WorkItem.
- * By default this is false, indicating that this counter is reported
- * as a delta.
- */
+
+ /// True if this counter is reported as the total cumulative aggregate
+ /// value accumulated since the worker started working on this WorkItem.
+ /// By default this is false, indicating that this counter is reported
+ /// as a delta.
core.bool cumulative;
- /** Distribution data */
+
+ /// Distribution data
DistributionUpdate distribution;
- /** Floating point value for Sum, Max, Min. */
+
+ /// Floating point value for Sum, Max, Min.
core.double floatingPoint;
- /** List of floating point numbers, for Set. */
+
+ /// List of floating point numbers, for Set.
FloatingPointList floatingPointList;
- /** Floating point mean aggregation value for Mean. */
+
+ /// Floating point mean aggregation value for Mean.
FloatingPointMean floatingPointMean;
- /** Integer value for Sum, Max, Min. */
+
+ /// Integer value for Sum, Max, Min.
SplitInt64 integer;
- /** List of integers, for Set. */
+
+ /// List of integers, for Set.
IntegerList integerList;
- /** Integer mean aggregation value for Mean. */
+
+ /// Integer mean aggregation value for Mean.
IntegerMean integerMean;
- /**
- * Value for internally-defined counters used by the Dataflow service.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Value for internally-defined counters used by the Dataflow service.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object internal;
- /** Counter name and aggregation type. */
+
+ /// Counter name and aggregation type.
NameAndKind nameAndKind;
- /**
- * The service-generated short identifier for this counter.
- * The short_id -> (name, metadata) mapping is constant for the lifetime of
- * a job.
- */
+
+ /// The service-generated short identifier for this counter.
+ /// The short_id -> (name, metadata) mapping is constant for the lifetime of
+ /// a job.
core.String shortId;
- /** List of strings, for Set. */
+
+ /// List of strings, for Set.
StringList stringList;
- /** Counter structured name and metadata. */
+
+ /// Counter structured name and metadata.
CounterStructuredNameAndMetadata structuredNameAndMetadata;
CounterUpdate();
@@ -2670,10 +2803,12 @@ class CounterUpdate {
floatingPoint = _json["floatingPoint"];
}
if (_json.containsKey("floatingPointList")) {
- floatingPointList = new FloatingPointList.fromJson(_json["floatingPointList"]);
+ floatingPointList =
+ new FloatingPointList.fromJson(_json["floatingPointList"]);
}
if (_json.containsKey("floatingPointMean")) {
- floatingPointMean = new FloatingPointMean.fromJson(_json["floatingPointMean"]);
+ floatingPointMean =
+ new FloatingPointMean.fromJson(_json["floatingPointMean"]);
}
if (_json.containsKey("integer")) {
integer = new SplitInt64.fromJson(_json["integer"]);
@@ -2697,12 +2832,14 @@ class CounterUpdate {
stringList = new StringList.fromJson(_json["stringList"]);
}
if (_json.containsKey("structuredNameAndMetadata")) {
- structuredNameAndMetadata = new CounterStructuredNameAndMetadata.fromJson(_json["structuredNameAndMetadata"]);
+ structuredNameAndMetadata = new CounterStructuredNameAndMetadata.fromJson(
+ _json["structuredNameAndMetadata"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (boolean != null) {
_json["boolean"] = boolean;
}
@@ -2749,21 +2886,23 @@ class CounterUpdate {
}
}
-/** A request to create a Cloud Dataflow job from a template. */
+/// A request to create a Cloud Dataflow job from a template.
class CreateJobFromTemplateRequest {
- /** The runtime environment for the job. */
+ /// The runtime environment for the job.
RuntimeEnvironment environment;
- /**
- * Required. A Cloud Storage path to the template from which to
- * create the job.
- * Must be a valid Cloud Storage URL, beginning with `gs://`.
- */
+
+ /// Required. A Cloud Storage path to the template from which to
+ /// create the job.
+ /// Must be a valid Cloud Storage URL, beginning with `gs://`.
core.String gcsPath;
- /** Required. The job name to use for the created job. */
+
+ /// Required. The job name to use for the created job.
core.String jobName;
- /** The location to which to direct the request. */
+
+ /// The location to which to direct the request.
core.String location;
- /** The runtime parameters to pass to the job. */
+
+ /// The runtime parameters to pass to the job.
core.Map<core.String, core.String> parameters;
CreateJobFromTemplateRequest();
@@ -2787,7 +2926,8 @@ class CreateJobFromTemplateRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (environment != null) {
_json["environment"] = (environment).toJson();
}
@@ -2807,9 +2947,9 @@ class CreateJobFromTemplateRequest {
}
}
-/** Identifies the location of a custom souce. */
+/// Identifies the location of a custom souce.
class CustomSourceLocation {
- /** Whether this source is stateful. */
+ /// Whether this source is stateful.
core.bool stateful;
CustomSourceLocation();
@@ -2821,7 +2961,8 @@ class CustomSourceLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (stateful != null) {
_json["stateful"] = stateful;
}
@@ -2829,19 +2970,16 @@ class CustomSourceLocation {
}
}
-/** Data disk assignment for a given VM instance. */
+/// Data disk assignment for a given VM instance.
class DataDiskAssignment {
- /**
- * Mounted data disks. The order is important a data disk's 0-based index in
- * this list defines which persistent directory the disk is mounted to, for
- * example the list of { "myproject-1014-104817-4c2-harness-0-disk-0" },
- * { "myproject-1014-104817-4c2-harness-0-disk-1" }.
- */
+ /// Mounted data disks. The order is important a data disk's 0-based index in
+ /// this list defines which persistent directory the disk is mounted to, for
+ /// example the list of { "myproject-1014-104817-4c2-harness-0-disk-0" },
+ /// { "myproject-1014-104817-4c2-harness-0-disk-1" }.
core.List<core.String> dataDisks;
- /**
- * VM instance name the data disks mounted to, for example
- * "myproject-1014-104817-4c2-harness-0".
- */
+
+ /// VM instance name the data disks mounted to, for example
+ /// "myproject-1014-104817-4c2-harness-0".
core.String vmInstance;
DataDiskAssignment();
@@ -2856,7 +2994,8 @@ class DataDiskAssignment {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dataDisks != null) {
_json["dataDisks"] = dataDisks;
}
@@ -2867,27 +3006,24 @@ class DataDiskAssignment {
}
}
-/**
- * Specification of one of the bundles produced as a result of splitting
- * a Source (e.g. when executing a SourceSplitRequest, or when
- * splitting an active task using WorkItemStatus.dynamic_source_split),
- * relative to the source being split.
- */
+/// Specification of one of the bundles produced as a result of splitting
+/// a Source (e.g. when executing a SourceSplitRequest, or when
+/// splitting an active task using WorkItemStatus.dynamic_source_split),
+/// relative to the source being split.
class DerivedSource {
- /**
- * What source to base the produced source on (if any).
- * Possible string values are:
- * - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or
- * unspecified.
- * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent
- * Source with no base.
- * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on the
- * Source being split.
- * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on
- * the base of the Source being split.
- */
+ /// What source to base the produced source on (if any).
+ /// Possible string values are:
+ /// - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or
+ /// unspecified.
+ /// - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent
+ /// Source with no base.
+ /// - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on
+ /// the Source being split.
+ /// - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on
+ /// the base of the Source being split.
core.String derivationMode;
- /** Specification of the source. */
+
+ /// Specification of the source.
Source source;
DerivedSource();
@@ -2902,7 +3038,8 @@ class DerivedSource {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (derivationMode != null) {
_json["derivationMode"] = derivationMode;
}
@@ -2913,35 +3050,33 @@ class DerivedSource {
}
}
-/** Describes the data disk used by a workflow job. */
+/// Describes the data disk used by a workflow job.
class Disk {
- /**
- * Disk storage type, as defined by Google Compute Engine. This
- * must be a disk type appropriate to the project and zone in which
- * the workers will run. If unknown or unspecified, the service
- * will attempt to choose a reasonable default.
- *
- * For example, the standard persistent disk type is a resource name
- * typically ending in "pd-standard". If SSD persistent disks are
- * available, the resource name typically ends with "pd-ssd". The
- * actual valid values are defined the Google Compute Engine API,
- * not by the Cloud Dataflow API; consult the Google Compute Engine
- * documentation for more information about determining the set of
- * available disk types for a particular project and zone.
- *
- * Google Compute Engine Disk types are local to a particular
- * project in a particular zone, and so the resource name will
- * typically look something like this:
- *
- * compute.googleapis.com/projects/project-id/zones/zone/diskTypes/pd-standard
- */
+ /// Disk storage type, as defined by Google Compute Engine. This
+ /// must be a disk type appropriate to the project and zone in which
+ /// the workers will run. If unknown or unspecified, the service
+ /// will attempt to choose a reasonable default.
+ ///
+ /// For example, the standard persistent disk type is a resource name
+ /// typically ending in "pd-standard". If SSD persistent disks are
+ /// available, the resource name typically ends with "pd-ssd". The
+ /// actual valid values are defined the Google Compute Engine API,
+ /// not by the Cloud Dataflow API; consult the Google Compute Engine
+ /// documentation for more information about determining the set of
+ /// available disk types for a particular project and zone.
+ ///
+ /// Google Compute Engine Disk types are local to a particular
+ /// project in a particular zone, and so the resource name will
+ /// typically look something like this:
+ ///
+ /// compute.googleapis.com/projects/project-id/zones/zone/diskTypes/pd-standard
core.String diskType;
- /** Directory in a VM where disk is mounted. */
+
+ /// Directory in a VM where disk is mounted.
core.String mountPoint;
- /**
- * Size of disk in GB. If zero or unspecified, the service will
- * attempt to choose a reasonable default.
- */
+
+ /// Size of disk in GB. If zero or unspecified, the service will
+ /// attempt to choose a reasonable default.
core.int sizeGb;
Disk();
@@ -2959,7 +3094,8 @@ class Disk {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (diskType != null) {
_json["diskType"] = diskType;
}
@@ -2973,47 +3109,52 @@ class Disk {
}
}
-/** Data provided with a pipeline or transform to provide descriptive info. */
+/// Data provided with a pipeline or transform to provide descriptive info.
class DisplayData {
- /** Contains value if the data is of a boolean type. */
+ /// Contains value if the data is of a boolean type.
core.bool boolValue;
- /** Contains value if the data is of duration type. */
+
+ /// Contains value if the data is of duration type.
core.String durationValue;
- /** Contains value if the data is of float type. */
+
+ /// Contains value if the data is of float type.
core.double floatValue;
- /** Contains value if the data is of int64 type. */
+
+ /// Contains value if the data is of int64 type.
core.String int64Value;
- /** Contains value if the data is of java class type. */
+
+ /// Contains value if the data is of java class type.
core.String javaClassValue;
- /**
- * The key identifying the display data.
- * This is intended to be used as a label for the display data
- * when viewed in a dax monitoring system.
- */
+
+ /// The key identifying the display data.
+ /// This is intended to be used as a label for the display data
+ /// when viewed in a dax monitoring system.
core.String key;
- /** An optional label to display in a dax UI for the element. */
+
+ /// An optional label to display in a dax UI for the element.
core.String label;
- /**
- * The namespace for the key. This is usually a class name or programming
- * language namespace (i.e. python module) which defines the display data.
- * This allows a dax monitoring system to specially handle the data
- * and perform custom rendering.
- */
+
+ /// The namespace for the key. This is usually a class name or programming
+ /// language namespace (i.e. python module) which defines the display data.
+ /// This allows a dax monitoring system to specially handle the data
+ /// and perform custom rendering.
core.String namespace;
- /**
- * A possible additional shorter value to display.
- * For example a java_class_name_value of com.mypackage.MyDoFn
- * will be stored with MyDoFn as the short_str_value and
- * com.mypackage.MyDoFn as the java_class_name value.
- * short_str_value can be displayed and java_class_name_value
- * will be displayed as a tooltip.
- */
+
+ /// A possible additional shorter value to display.
+ /// For example a java_class_name_value of com.mypackage.MyDoFn
+ /// will be stored with MyDoFn as the short_str_value and
+ /// com.mypackage.MyDoFn as the java_class_name value.
+ /// short_str_value can be displayed and java_class_name_value
+ /// will be displayed as a tooltip.
core.String shortStrValue;
- /** Contains value if the data is of string type. */
+
+ /// Contains value if the data is of string type.
core.String strValue;
- /** Contains value if the data is of timestamp type. */
+
+ /// Contains value if the data is of timestamp type.
core.String timestampValue;
- /** An optional full URL. */
+
+ /// An optional full URL.
core.String url;
DisplayData();
@@ -3058,7 +3199,8 @@ class DisplayData {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (boolValue != null) {
_json["boolValue"] = boolValue;
}
@@ -3099,25 +3241,27 @@ class DisplayData {
}
}
-/** A metric value representing a distribution. */
+/// A metric value representing a distribution.
class DistributionUpdate {
- /** The count of the number of elements present in the distribution. */
+ /// The count of the number of elements present in the distribution.
SplitInt64 count;
- /**
- * (Optional) Logarithmic histogram of values.
- * Each log may be in no more than one bucket. Order does not matter.
- */
+
+ /// (Optional) Logarithmic histogram of values.
+ /// Each log may be in no more than one bucket. Order does not matter.
core.List<LogBucket> logBuckets;
- /** The maximum value present in the distribution. */
+
+ /// The maximum value present in the distribution.
SplitInt64 max;
- /** The minimum value present in the distribution. */
+
+ /// The minimum value present in the distribution.
SplitInt64 min;
- /**
- * Use an int64 since we'd prefer the added precision. If overflow is a common
- * problem we can detect it and use an additional int64 or a double.
- */
+
+ /// Use an int64 since we'd prefer the added precision. If overflow is a
+ /// common
+ /// problem we can detect it and use an additional int64 or a double.
SplitInt64 sum;
- /** Use a double since the sum of squares is likely to overflow int64. */
+
+ /// Use a double since the sum of squares is likely to overflow int64.
core.double sumOfSquares;
DistributionUpdate();
@@ -3127,7 +3271,9 @@ class DistributionUpdate {
count = new SplitInt64.fromJson(_json["count"]);
}
if (_json.containsKey("logBuckets")) {
- logBuckets = _json["logBuckets"].map((value) => new LogBucket.fromJson(value)).toList();
+ logBuckets = _json["logBuckets"]
+ .map((value) => new LogBucket.fromJson(value))
+ .toList();
}
if (_json.containsKey("max")) {
max = new SplitInt64.fromJson(_json["max"]);
@@ -3144,12 +3290,14 @@ class DistributionUpdate {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (count != null) {
_json["count"] = (count).toJson();
}
if (logBuckets != null) {
- _json["logBuckets"] = logBuckets.map((value) => (value).toJson()).toList();
+ _json["logBuckets"] =
+ logBuckets.map((value) => (value).toJson()).toList();
}
if (max != null) {
_json["max"] = (max).toJson();
@@ -3167,22 +3315,17 @@ class DistributionUpdate {
}
}
-/**
- * When a task splits using WorkItemStatus.dynamic_source_split, this
- * message describes the two parts of the split relative to the
- * description of the current task's input.
- */
+/// When a task splits using WorkItemStatus.dynamic_source_split, this
+/// message describes the two parts of the split relative to the
+/// description of the current task's input.
class DynamicSourceSplit {
- /**
- * Primary part (continued to be processed by worker).
- * Specified relative to the previously-current source.
- * Becomes current.
- */
+ /// Primary part (continued to be processed by worker).
+ /// Specified relative to the previously-current source.
+ /// Becomes current.
DerivedSource primary;
- /**
- * Residual part (returned to the pool of work).
- * Specified relative to the previously-current source.
- */
+
+ /// Residual part (returned to the pool of work).
+ /// Specified relative to the previously-current source.
DerivedSource residual;
DynamicSourceSplit();
@@ -3197,7 +3340,8 @@ class DynamicSourceSplit {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (primary != null) {
_json["primary"] = (primary).toJson();
}
@@ -3208,81 +3352,74 @@ class DynamicSourceSplit {
}
}
-/** Describes the environment in which a Dataflow Job runs. */
+/// Describes the environment in which a Dataflow Job runs.
class Environment {
- /**
- * The type of cluster manager API to use. If unknown or
- * unspecified, the service will attempt to choose a reasonable
- * default. This should be in the form of the API service name,
- * e.g. "compute.googleapis.com".
- */
+ /// The type of cluster manager API to use. If unknown or
+ /// unspecified, the service will attempt to choose a reasonable
+ /// default. This should be in the form of the API service name,
+ /// e.g. "compute.googleapis.com".
core.String clusterManagerApiService;
- /**
- * The dataset for the current project where various workflow
- * related tables are stored.
- *
- * The supported resource type is:
- *
- * Google BigQuery:
- * bigquery.googleapis.com/{dataset}
- */
+
+ /// The dataset for the current project where various workflow
+ /// related tables are stored.
+ ///
+ /// The supported resource type is:
+ ///
+ /// Google BigQuery:
+ /// bigquery.googleapis.com/{dataset}
core.String dataset;
- /** The list of experiments to enable. */
+
+ /// The list of experiments to enable.
core.List<core.String> experiments;
- /**
- * Experimental settings.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Experimental settings.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> internalExperiments;
- /**
- * The Cloud Dataflow SDK pipeline options specified by the user. These
- * options are passed through the service and are used to recreate the
- * SDK pipeline options on the worker in a language agnostic and platform
- * independent way.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The Cloud Dataflow SDK pipeline options specified by the user. These
+ /// options are passed through the service and are used to recreate the
+ /// SDK pipeline options on the worker in a language agnostic and platform
+ /// independent way.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> sdkPipelineOptions;
- /** Identity to run virtual machines as. Defaults to the default account. */
+
+ /// Identity to run virtual machines as. Defaults to the default account.
core.String serviceAccountEmail;
- /**
- * The prefix of the resources the system should use for temporary
- * storage. The system will append the suffix "/temp-{JOBNAME} to
- * this resource prefix, where {JOBNAME} is the value of the
- * job_name field. The resulting bucket and object prefix is used
- * as the prefix of the resources used to store temporary data
- * needed during the job execution. NOTE: This will override the
- * value in taskrunner_settings.
- * The supported resource type is:
- *
- * Google Cloud Storage:
- *
- * storage.googleapis.com/{bucket}/{object}
- * bucket.storage.googleapis.com/{object}
- */
+
+ /// The prefix of the resources the system should use for temporary
+ /// storage. The system will append the suffix "/temp-{JOBNAME} to
+ /// this resource prefix, where {JOBNAME} is the value of the
+ /// job_name field. The resulting bucket and object prefix is used
+ /// as the prefix of the resources used to store temporary data
+ /// needed during the job execution. NOTE: This will override the
+ /// value in taskrunner_settings.
+ /// The supported resource type is:
+ ///
+ /// Google Cloud Storage:
+ ///
+ /// storage.googleapis.com/{bucket}/{object}
+ /// bucket.storage.googleapis.com/{object}
core.String tempStoragePrefix;
- /**
- * A description of the process that generated the request.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A description of the process that generated the request.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> userAgent;
- /**
- * A structure describing which components and their versions of the service
- * are required in order to run the job.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A structure describing which components and their versions of the service
+ /// are required in order to run the job.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> version;
- /**
- * The worker pools. At least one "harness" worker pool must be
- * specified in order for the job to have workers.
- */
+
+ /// The worker pools. At least one "harness" worker pool must be
+ /// specified in order for the job to have workers.
core.List<WorkerPool> workerPools;
Environment();
@@ -3316,12 +3453,15 @@ class Environment {
version = _json["version"];
}
if (_json.containsKey("workerPools")) {
- workerPools = _json["workerPools"].map((value) => new WorkerPool.fromJson(value)).toList();
+ workerPools = _json["workerPools"]
+ .map((value) => new WorkerPool.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (clusterManagerApiService != null) {
_json["clusterManagerApiService"] = clusterManagerApiService;
}
@@ -3350,69 +3490,77 @@ class Environment {
_json["version"] = version;
}
if (workerPools != null) {
- _json["workerPools"] = workerPools.map((value) => (value).toJson()).toList();
+ _json["workerPools"] =
+ workerPools.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** A message describing the state of a particular execution stage. */
+/// A message describing the state of a particular execution stage.
class ExecutionStageState {
- /** The time at which the stage transitioned to this state. */
+ /// The time at which the stage transitioned to this state.
core.String currentStateTime;
- /** The name of the execution stage. */
+
+ /// The name of the execution stage.
core.String executionStageName;
- /**
- * Executions stage states allow the same set of values as JobState.
- * Possible string values are:
- * - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
- * - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has not
- * yet started to run.
- * - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
- * currently running.
- * - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
- * successfully completed.
- * This is a terminal job state. This state may be set by the Cloud Dataflow
- * service, as a transition from `JOB_STATE_RUNNING`. It may also be set via a
- * Cloud Dataflow `UpdateJob` call, if the job has not yet reached a terminal
- * state.
- * - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
- * failed. This is a
- * terminal job state. This state may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_RUNNING`.
- * - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job has
- * been explicitly
- * cancelled. This is a terminal job state. This state may only be
- * set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
- * yet reached another terminal state.
- * - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
- * successfully updated,
- * meaning that this job was stopped and another job was started, inheriting
- * state from this one. This is a terminal job state. This state may only be
- * set by the Cloud Dataflow service, and only as a transition from
- * `JOB_STATE_RUNNING`.
- * - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is in
- * the process of draining.
- * A draining job has stopped pulling from its input sources and is processing
- * any data that remains in-flight. This state may be set via a Cloud Dataflow
- * `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
- * that are draining may only transition to `JOB_STATE_DRAINED`,
- * `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
- * - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has been
- * drained.
- * A drained job terminated by stopping pulling from its input sources and
- * processing any data that remained in-flight when draining was requested.
- * This state is a terminal state, may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_DRAINING`.
- * - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has been
- * created but is not yet
- * running. Jobs that are pending may only transition to `JOB_STATE_RUNNING`,
- * or `JOB_STATE_FAILED`.
- * - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
- * has been explicitly cancelled
- * and is in the process of stopping. Jobs that are cancelling may only
- * transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
- */
+
+ /// Executions stage states allow the same set of values as JobState.
+ /// Possible string values are:
+ /// - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
+ /// - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has
+ /// not
+ /// yet started to run.
+ /// - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
+ /// currently running.
+ /// - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
+ /// successfully completed.
+ /// This is a terminal job state. This state may be set by the Cloud
+ /// Dataflow
+ /// service, as a transition from `JOB_STATE_RUNNING`. It may also be set via
+ /// a
+ /// Cloud Dataflow `UpdateJob` call, if the job has not yet reached a
+ /// terminal
+ /// state.
+ /// - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
+ /// failed. This is a
+ /// terminal job state. This state may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job
+ /// has been explicitly
+ /// cancelled. This is a terminal job state. This state may only be
+ /// set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
+ /// yet reached another terminal state.
+ /// - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
+ /// successfully updated,
+ /// meaning that this job was stopped and another job was started, inheriting
+ /// state from this one. This is a terminal job state. This state may only be
+ /// set by the Cloud Dataflow service, and only as a transition from
+ /// `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is
+ /// in the process of draining.
+ /// A draining job has stopped pulling from its input sources and is
+ /// processing
+ /// any data that remains in-flight. This state may be set via a Cloud
+ /// Dataflow
+ /// `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
+ /// that are draining may only transition to `JOB_STATE_DRAINED`,
+ /// `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has
+ /// been drained.
+ /// A drained job terminated by stopping pulling from its input sources and
+ /// processing any data that remained in-flight when draining was requested.
+ /// This state is a terminal state, may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_DRAINING`.
+ /// - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has
+ /// been created but is not yet
+ /// running. Jobs that are pending may only transition to
+ /// `JOB_STATE_RUNNING`,
+ /// or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
+ /// has been explicitly cancelled
+ /// and is in the process of stopping. Jobs that are cancelling may only
+ /// transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
core.String executionStageState;
ExecutionStageState();
@@ -3430,7 +3578,8 @@ class ExecutionStageState {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (currentStateTime != null) {
_json["currentStateTime"] = currentStateTime;
}
@@ -3444,57 +3593,63 @@ class ExecutionStageState {
}
}
-/**
- * Description of the composing transforms, names/ids, and input/outputs of a
- * stage of execution. Some composing transforms and sources may have been
- * generated by the Dataflow service during execution planning.
- */
+/// Description of the composing transforms, names/ids, and input/outputs of a
+/// stage of execution. Some composing transforms and sources may have been
+/// generated by the Dataflow service during execution planning.
class ExecutionStageSummary {
- /**
- * Collections produced and consumed by component transforms of this stage.
- */
+ /// Collections produced and consumed by component transforms of this stage.
core.List<ComponentSource> componentSource;
- /** Transforms that comprise this execution stage. */
+
+ /// Transforms that comprise this execution stage.
core.List<ComponentTransform> componentTransform;
- /** Dataflow service generated id for this stage. */
+
+ /// Dataflow service generated id for this stage.
core.String id;
- /** Input sources for this stage. */
+
+ /// Input sources for this stage.
core.List<StageSource> inputSource;
- /**
- * Type of tranform this stage is executing.
- * Possible string values are:
- * - "UNKNOWN_KIND" : Unrecognized transform type.
- * - "PAR_DO_KIND" : ParDo transform.
- * - "GROUP_BY_KEY_KIND" : Group By Key transform.
- * - "FLATTEN_KIND" : Flatten transform.
- * - "READ_KIND" : Read transform.
- * - "WRITE_KIND" : Write transform.
- * - "CONSTANT_KIND" : Constructs from a constant value, such as with
- * Create.of.
- * - "SINGLETON_KIND" : Creates a Singleton view of a collection.
- * - "SHUFFLE_KIND" : Opening or closing a shuffle session, often as part of a
- * GroupByKey.
- */
+
+ /// Type of tranform this stage is executing.
+ /// Possible string values are:
+ /// - "UNKNOWN_KIND" : Unrecognized transform type.
+ /// - "PAR_DO_KIND" : ParDo transform.
+ /// - "GROUP_BY_KEY_KIND" : Group By Key transform.
+ /// - "FLATTEN_KIND" : Flatten transform.
+ /// - "READ_KIND" : Read transform.
+ /// - "WRITE_KIND" : Write transform.
+ /// - "CONSTANT_KIND" : Constructs from a constant value, such as with
+ /// Create.of.
+ /// - "SINGLETON_KIND" : Creates a Singleton view of a collection.
+ /// - "SHUFFLE_KIND" : Opening or closing a shuffle session, often as part of
+ /// a GroupByKey.
core.String kind;
- /** Dataflow service generated name for this stage. */
+
+ /// Dataflow service generated name for this stage.
core.String name;
- /** Output sources for this stage. */
+
+ /// Output sources for this stage.
core.List<StageSource> outputSource;
ExecutionStageSummary();
ExecutionStageSummary.fromJson(core.Map _json) {
if (_json.containsKey("componentSource")) {
- componentSource = _json["componentSource"].map((value) => new ComponentSource.fromJson(value)).toList();
+ componentSource = _json["componentSource"]
+ .map((value) => new ComponentSource.fromJson(value))
+ .toList();
}
if (_json.containsKey("componentTransform")) {
- componentTransform = _json["componentTransform"].map((value) => new ComponentTransform.fromJson(value)).toList();
+ componentTransform = _json["componentTransform"]
+ .map((value) => new ComponentTransform.fromJson(value))
+ .toList();
}
if (_json.containsKey("id")) {
id = _json["id"];
}
if (_json.containsKey("inputSource")) {
- inputSource = _json["inputSource"].map((value) => new StageSource.fromJson(value)).toList();
+ inputSource = _json["inputSource"]
+ .map((value) => new StageSource.fromJson(value))
+ .toList();
}
if (_json.containsKey("kind")) {
kind = _json["kind"];
@@ -3503,23 +3658,29 @@ class ExecutionStageSummary {
name = _json["name"];
}
if (_json.containsKey("outputSource")) {
- outputSource = _json["outputSource"].map((value) => new StageSource.fromJson(value)).toList();
+ outputSource = _json["outputSource"]
+ .map((value) => new StageSource.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (componentSource != null) {
- _json["componentSource"] = componentSource.map((value) => (value).toJson()).toList();
+ _json["componentSource"] =
+ componentSource.map((value) => (value).toJson()).toList();
}
if (componentTransform != null) {
- _json["componentTransform"] = componentTransform.map((value) => (value).toJson()).toList();
+ _json["componentTransform"] =
+ componentTransform.map((value) => (value).toJson()).toList();
}
if (id != null) {
_json["id"] = id;
}
if (inputSource != null) {
- _json["inputSource"] = inputSource.map((value) => (value).toJson()).toList();
+ _json["inputSource"] =
+ inputSource.map((value) => (value).toJson()).toList();
}
if (kind != null) {
_json["kind"] = kind;
@@ -3528,15 +3689,16 @@ class ExecutionStageSummary {
_json["name"] = name;
}
if (outputSource != null) {
- _json["outputSource"] = outputSource.map((value) => (value).toJson()).toList();
+ _json["outputSource"] =
+ outputSource.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** Indicates which location failed to respond to a request for data. */
+/// Indicates which location failed to respond to a request for data.
class FailedLocation {
- /** The name of the failed location. */
+ /// The name of the failed location.
core.String name;
FailedLocation();
@@ -3548,7 +3710,8 @@ class FailedLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (name != null) {
_json["name"] = name;
}
@@ -3556,23 +3719,25 @@ class FailedLocation {
}
}
-/**
- * An instruction that copies its inputs (zero or more) to its (single) output.
- */
+/// An instruction that copies its inputs (zero or more) to its (single)
+/// output.
class FlattenInstruction {
- /** Describes the inputs to the flatten instruction. */
+ /// Describes the inputs to the flatten instruction.
core.List<InstructionInput> inputs;
FlattenInstruction();
FlattenInstruction.fromJson(core.Map _json) {
if (_json.containsKey("inputs")) {
- inputs = _json["inputs"].map((value) => new InstructionInput.fromJson(value)).toList();
+ inputs = _json["inputs"]
+ .map((value) => new InstructionInput.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (inputs != null) {
_json["inputs"] = inputs.map((value) => (value).toJson()).toList();
}
@@ -3580,9 +3745,9 @@ class FlattenInstruction {
}
}
-/** A metric value representing a list of floating point numbers. */
+/// A metric value representing a list of floating point numbers.
class FloatingPointList {
- /** Elements of the list. */
+ /// Elements of the list.
core.List<core.double> elements;
FloatingPointList();
@@ -3594,7 +3759,8 @@ class FloatingPointList {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (elements != null) {
_json["elements"] = elements;
}
@@ -3602,11 +3768,12 @@ class FloatingPointList {
}
}
-/** A representation of a floating point mean metric contribution. */
+/// A representation of a floating point mean metric contribution.
class FloatingPointMean {
- /** The number of values being aggregated. */
+ /// The number of values being aggregated.
SplitInt64 count;
- /** The sum of all values being aggregated. */
+
+ /// The sum of all values being aggregated.
core.double sum;
FloatingPointMean();
@@ -3621,7 +3788,8 @@ class FloatingPointMean {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (count != null) {
_json["count"] = (count).toJson();
}
@@ -3632,16 +3800,16 @@ class FloatingPointMean {
}
}
-/** Request to get updated debug configuration for component. */
+/// Request to get updated debug configuration for component.
class GetDebugConfigRequest {
- /**
- * The internal component id for which debug configuration is
- * requested.
- */
+ /// The internal component id for which debug configuration is
+ /// requested.
core.String componentId;
- /** The location which contains the job specified by job_id. */
+
+ /// The location which contains the job specified by job_id.
core.String location;
- /** The worker id, i.e., VM hostname. */
+
+ /// The worker id, i.e., VM hostname.
core.String workerId;
GetDebugConfigRequest();
@@ -3659,7 +3827,8 @@ class GetDebugConfigRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (componentId != null) {
_json["componentId"] = componentId;
}
@@ -3673,9 +3842,9 @@ class GetDebugConfigRequest {
}
}
-/** Response to a get debug configuration request. */
+/// Response to a get debug configuration request.
class GetDebugConfigResponse {
- /** The encoded debug configuration for the requested component. */
+ /// The encoded debug configuration for the requested component.
core.String config;
GetDebugConfigResponse();
@@ -3687,7 +3856,8 @@ class GetDebugConfigResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (config != null) {
_json["config"] = config;
}
@@ -3695,17 +3865,14 @@ class GetDebugConfigResponse {
}
}
-/** The response to a GetTemplate request. */
+/// The response to a GetTemplate request.
class GetTemplateResponse {
- /**
- * The template metadata describing the template name, available
- * parameters, etc.
- */
+ /// The template metadata describing the template name, available
+ /// parameters, etc.
TemplateMetadata metadata;
- /**
- * The status of the get template request. Any problems with the
- * request will be indicated in the error_details.
- */
+
+ /// The status of the get template request. Any problems with the
+ /// request will be indicated in the error_details.
Status status;
GetTemplateResponse();
@@ -3720,7 +3887,8 @@ class GetTemplateResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (metadata != null) {
_json["metadata"] = (metadata).toJson();
}
@@ -3731,19 +3899,16 @@ class GetTemplateResponse {
}
}
-/**
- * An input of an instruction, as a reference to an output of a
- * producer instruction.
- */
+/// An input of an instruction, as a reference to an output of a
+/// producer instruction.
class InstructionInput {
- /** The output index (origin zero) within the producer. */
+ /// The output index (origin zero) within the producer.
core.int outputNum;
- /**
- * The index (origin zero) of the parallel instruction that produces
- * the output to be consumed by this input. This index is relative
- * to the list of instructions in this input's instruction's
- * containing MapTask.
- */
+
+ /// The index (origin zero) of the parallel instruction that produces
+ /// the output to be consumed by this input. This index is relative
+ /// to the list of instructions in this input's instruction's
+ /// containing MapTask.
core.int producerInstructionIndex;
InstructionInput();
@@ -3758,7 +3923,8 @@ class InstructionInput {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (outputNum != null) {
_json["outputNum"] = outputNum;
}
@@ -3769,36 +3935,32 @@ class InstructionInput {
}
}
-/** An output of an instruction. */
+/// An output of an instruction.
class InstructionOutput {
- /**
- * The codec to use to encode data being written via this output.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// The codec to use to encode data being written via this output.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> codec;
- /** The user-provided name of this output. */
+
+ /// The user-provided name of this output.
core.String name;
- /**
- * For system-generated byte and mean byte metrics, certain instructions
- * should only report the key size.
- */
+
+ /// For system-generated byte and mean byte metrics, certain instructions
+ /// should only report the key size.
core.bool onlyCountKeyBytes;
- /**
- * For system-generated byte and mean byte metrics, certain instructions
- * should only report the value size.
- */
+
+ /// For system-generated byte and mean byte metrics, certain instructions
+ /// should only report the value size.
core.bool onlyCountValueBytes;
- /**
- * System-defined name for this output in the original workflow graph.
- * Outputs that do not contribute to an original instruction do not set this.
- */
+
+ /// System-defined name for this output in the original workflow graph.
+ /// Outputs that do not contribute to an original instruction do not set
+ /// this.
core.String originalName;
- /**
- * System-defined name of this output.
- * Unique across the workflow.
- */
+
+ /// System-defined name of this output.
+ /// Unique across the workflow.
core.String systemName;
InstructionOutput();
@@ -3825,7 +3987,8 @@ class InstructionOutput {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (codec != null) {
_json["codec"] = codec;
}
@@ -3848,21 +4011,24 @@ class InstructionOutput {
}
}
-/** A metric value representing a list of integers. */
+/// A metric value representing a list of integers.
class IntegerList {
- /** Elements of the list. */
+ /// Elements of the list.
core.List<SplitInt64> elements;
IntegerList();
IntegerList.fromJson(core.Map _json) {
if (_json.containsKey("elements")) {
- elements = _json["elements"].map((value) => new SplitInt64.fromJson(value)).toList();
+ elements = _json["elements"]
+ .map((value) => new SplitInt64.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (elements != null) {
_json["elements"] = elements.map((value) => (value).toJson()).toList();
}
@@ -3870,11 +4036,12 @@ class IntegerList {
}
}
-/** A representation of an integer mean metric contribution. */
+/// A representation of an integer mean metric contribution.
class IntegerMean {
- /** The number of values being aggregated. */
+ /// The number of values being aggregated.
SplitInt64 count;
- /** The sum of all values being aggregated. */
+
+ /// The sum of all values being aggregated.
SplitInt64 sum;
IntegerMean();
@@ -3889,7 +4056,8 @@ class IntegerMean {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (count != null) {
_json["count"] = (count).toJson();
}
@@ -3900,243 +4068,252 @@ class IntegerMean {
}
}
-/** Defines a job to be run by the Cloud Dataflow service. */
+/// Defines a job to be run by the Cloud Dataflow service.
class Job {
- /**
- * The client's unique identifier of the job, re-used across retried attempts.
- * If this field is set, the service will ensure its uniqueness.
- * The request to create a job will fail if the service has knowledge of a
- * previously submitted job with the same client's ID and job name.
- * The caller may use this field to ensure idempotence of job
- * creation across retried attempts to create a job.
- * By default, the field is empty and, in that case, the service ignores it.
- */
+ /// The client's unique identifier of the job, re-used across retried
+ /// attempts.
+ /// If this field is set, the service will ensure its uniqueness.
+ /// The request to create a job will fail if the service has knowledge of a
+ /// previously submitted job with the same client's ID and job name.
+ /// The caller may use this field to ensure idempotence of job
+ /// creation across retried attempts to create a job.
+ /// By default, the field is empty and, in that case, the service ignores it.
core.String clientRequestId;
- /**
- * The timestamp when the job was initially created. Immutable and set by the
- * Cloud Dataflow service.
- */
+
+ /// The timestamp when the job was initially created. Immutable and set by
+ /// the
+ /// Cloud Dataflow service.
core.String createTime;
- /**
- * The current state of the job.
- *
- * Jobs are created in the `JOB_STATE_STOPPED` state unless otherwise
- * specified.
- *
- * A job in the `JOB_STATE_RUNNING` state may asynchronously enter a
- * terminal state. After a job has reached a terminal state, no
- * further state updates may be made.
- *
- * This field may be mutated by the Cloud Dataflow service;
- * callers cannot mutate it.
- * Possible string values are:
- * - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
- * - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has not
- * yet started to run.
- * - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
- * currently running.
- * - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
- * successfully completed.
- * This is a terminal job state. This state may be set by the Cloud Dataflow
- * service, as a transition from `JOB_STATE_RUNNING`. It may also be set via a
- * Cloud Dataflow `UpdateJob` call, if the job has not yet reached a terminal
- * state.
- * - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
- * failed. This is a
- * terminal job state. This state may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_RUNNING`.
- * - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job has
- * been explicitly
- * cancelled. This is a terminal job state. This state may only be
- * set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
- * yet reached another terminal state.
- * - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
- * successfully updated,
- * meaning that this job was stopped and another job was started, inheriting
- * state from this one. This is a terminal job state. This state may only be
- * set by the Cloud Dataflow service, and only as a transition from
- * `JOB_STATE_RUNNING`.
- * - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is in
- * the process of draining.
- * A draining job has stopped pulling from its input sources and is processing
- * any data that remains in-flight. This state may be set via a Cloud Dataflow
- * `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
- * that are draining may only transition to `JOB_STATE_DRAINED`,
- * `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
- * - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has been
- * drained.
- * A drained job terminated by stopping pulling from its input sources and
- * processing any data that remained in-flight when draining was requested.
- * This state is a terminal state, may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_DRAINING`.
- * - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has been
- * created but is not yet
- * running. Jobs that are pending may only transition to `JOB_STATE_RUNNING`,
- * or `JOB_STATE_FAILED`.
- * - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
- * has been explicitly cancelled
- * and is in the process of stopping. Jobs that are cancelling may only
- * transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
- */
+
+ /// The current state of the job.
+ ///
+ /// Jobs are created in the `JOB_STATE_STOPPED` state unless otherwise
+ /// specified.
+ ///
+ /// A job in the `JOB_STATE_RUNNING` state may asynchronously enter a
+ /// terminal state. After a job has reached a terminal state, no
+ /// further state updates may be made.
+ ///
+ /// This field may be mutated by the Cloud Dataflow service;
+ /// callers cannot mutate it.
+ /// Possible string values are:
+ /// - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
+ /// - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has
+ /// not
+ /// yet started to run.
+ /// - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
+ /// currently running.
+ /// - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
+ /// successfully completed.
+ /// This is a terminal job state. This state may be set by the Cloud
+ /// Dataflow
+ /// service, as a transition from `JOB_STATE_RUNNING`. It may also be set via
+ /// a
+ /// Cloud Dataflow `UpdateJob` call, if the job has not yet reached a
+ /// terminal
+ /// state.
+ /// - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
+ /// failed. This is a
+ /// terminal job state. This state may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job
+ /// has been explicitly
+ /// cancelled. This is a terminal job state. This state may only be
+ /// set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
+ /// yet reached another terminal state.
+ /// - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
+ /// successfully updated,
+ /// meaning that this job was stopped and another job was started, inheriting
+ /// state from this one. This is a terminal job state. This state may only be
+ /// set by the Cloud Dataflow service, and only as a transition from
+ /// `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is
+ /// in the process of draining.
+ /// A draining job has stopped pulling from its input sources and is
+ /// processing
+ /// any data that remains in-flight. This state may be set via a Cloud
+ /// Dataflow
+ /// `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
+ /// that are draining may only transition to `JOB_STATE_DRAINED`,
+ /// `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has
+ /// been drained.
+ /// A drained job terminated by stopping pulling from its input sources and
+ /// processing any data that remained in-flight when draining was requested.
+ /// This state is a terminal state, may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_DRAINING`.
+ /// - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has
+ /// been created but is not yet
+ /// running. Jobs that are pending may only transition to
+ /// `JOB_STATE_RUNNING`,
+ /// or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
+ /// has been explicitly cancelled
+ /// and is in the process of stopping. Jobs that are cancelling may only
+ /// transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
core.String currentState;
- /** The timestamp associated with the current state. */
+
+ /// The timestamp associated with the current state.
core.String currentStateTime;
- /** The environment for the job. */
+
+ /// The environment for the job.
Environment environment;
- /** Deprecated. */
+
+ /// Deprecated.
JobExecutionInfo executionInfo;
- /**
- * The unique ID of this job.
- *
- * This field is set by the Cloud Dataflow service when the Job is
- * created, and is immutable for the life of the job.
- */
+
+ /// The unique ID of this job.
+ ///
+ /// This field is set by the Cloud Dataflow service when the Job is
+ /// created, and is immutable for the life of the job.
core.String id;
- /**
- * User-defined labels for this job.
- *
- * The labels map can contain no more than 64 entries. Entries of the labels
- * map are UTF8 strings that comply with the following restrictions:
- *
- * * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62}
- * * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63}
- * * Both keys and values are additionally constrained to be <= 128 bytes in
- * size.
- */
+
+ /// User-defined labels for this job.
+ ///
+ /// The labels map can contain no more than 64 entries. Entries of the
+ /// labels
+ /// map are UTF8 strings that comply with the following restrictions:
+ ///
+ /// * Keys must conform to regexp: \p{Ll}\p{Lo}{0,62}
+ /// * Values must conform to regexp: [\p{Ll}\p{Lo}\p{N}_-]{0,63}
+ /// * Both keys and values are additionally constrained to be <= 128 bytes in
+ /// size.
core.Map<core.String, core.String> labels;
- /** The location that contains this job. */
+
+ /// The location that contains this job.
core.String location;
- /**
- * The user-specified Cloud Dataflow job name.
- *
- * Only one Job with a given name may exist in a project at any
- * given time. If a caller attempts to create a Job with the same
- * name as an already-existing Job, the attempt returns the
- * existing Job.
- *
- * The name must match the regular expression
- * `[a-z]([-a-z0-9]{0,38}[a-z0-9])?`
- */
+
+ /// The user-specified Cloud Dataflow job name.
+ ///
+ /// Only one Job with a given name may exist in a project at any
+ /// given time. If a caller attempts to create a Job with the same
+ /// name as an already-existing Job, the attempt returns the
+ /// existing Job.
+ ///
+ /// The name must match the regular expression
+ /// `[a-z]([-a-z0-9]{0,38}[a-z0-9])?`
core.String name;
- /**
- * Preliminary field: The format of this data may change at any time.
- * A description of the user pipeline and stages through which it is executed.
- * Created by Cloud Dataflow service. Only retrieved with
- * JOB_VIEW_DESCRIPTION or JOB_VIEW_ALL.
- */
+
+ /// Preliminary field: The format of this data may change at any time.
+ /// A description of the user pipeline and stages through which it is
+ /// executed.
+ /// Created by Cloud Dataflow service. Only retrieved with
+ /// JOB_VIEW_DESCRIPTION or JOB_VIEW_ALL.
PipelineDescription pipelineDescription;
- /** The ID of the Cloud Platform project that the job belongs to. */
+
+ /// The ID of the Cloud Platform project that the job belongs to.
core.String projectId;
- /**
- * If this job is an update of an existing job, this field is the job ID
- * of the job it replaced.
- *
- * When sending a `CreateJobRequest`, you can update a job by specifying it
- * here. The job named here is stopped, and its intermediate state is
- * transferred to this job.
- */
+
+ /// If this job is an update of an existing job, this field is the job ID
+ /// of the job it replaced.
+ ///
+ /// When sending a `CreateJobRequest`, you can update a job by specifying it
+ /// here. The job named here is stopped, and its intermediate state is
+ /// transferred to this job.
core.String replaceJobId;
- /**
- * If another job is an update of this job (and thus, this job is in
- * `JOB_STATE_UPDATED`), this field contains the ID of that job.
- */
+
+ /// If another job is an update of this job (and thus, this job is in
+ /// `JOB_STATE_UPDATED`), this field contains the ID of that job.
core.String replacedByJobId;
- /**
- * The job's requested state.
- *
- * `UpdateJob` may be used to switch between the `JOB_STATE_STOPPED` and
- * `JOB_STATE_RUNNING` states, by setting requested_state. `UpdateJob` may
- * also be used to directly set a job's requested state to
- * `JOB_STATE_CANCELLED` or `JOB_STATE_DONE`, irrevocably terminating the
- * job if it has not already reached a terminal state.
- * Possible string values are:
- * - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
- * - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has not
- * yet started to run.
- * - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
- * currently running.
- * - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
- * successfully completed.
- * This is a terminal job state. This state may be set by the Cloud Dataflow
- * service, as a transition from `JOB_STATE_RUNNING`. It may also be set via a
- * Cloud Dataflow `UpdateJob` call, if the job has not yet reached a terminal
- * state.
- * - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
- * failed. This is a
- * terminal job state. This state may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_RUNNING`.
- * - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job has
- * been explicitly
- * cancelled. This is a terminal job state. This state may only be
- * set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
- * yet reached another terminal state.
- * - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
- * successfully updated,
- * meaning that this job was stopped and another job was started, inheriting
- * state from this one. This is a terminal job state. This state may only be
- * set by the Cloud Dataflow service, and only as a transition from
- * `JOB_STATE_RUNNING`.
- * - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is in
- * the process of draining.
- * A draining job has stopped pulling from its input sources and is processing
- * any data that remains in-flight. This state may be set via a Cloud Dataflow
- * `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
- * that are draining may only transition to `JOB_STATE_DRAINED`,
- * `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
- * - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has been
- * drained.
- * A drained job terminated by stopping pulling from its input sources and
- * processing any data that remained in-flight when draining was requested.
- * This state is a terminal state, may only be set by the Cloud Dataflow
- * service, and only as a transition from `JOB_STATE_DRAINING`.
- * - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has been
- * created but is not yet
- * running. Jobs that are pending may only transition to `JOB_STATE_RUNNING`,
- * or `JOB_STATE_FAILED`.
- * - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
- * has been explicitly cancelled
- * and is in the process of stopping. Jobs that are cancelling may only
- * transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
- */
+
+ /// The job's requested state.
+ ///
+ /// `UpdateJob` may be used to switch between the `JOB_STATE_STOPPED` and
+ /// `JOB_STATE_RUNNING` states, by setting requested_state. `UpdateJob` may
+ /// also be used to directly set a job's requested state to
+ /// `JOB_STATE_CANCELLED` or `JOB_STATE_DONE`, irrevocably terminating the
+ /// job if it has not already reached a terminal state.
+ /// Possible string values are:
+ /// - "JOB_STATE_UNKNOWN" : The job's run state isn't specified.
+ /// - "JOB_STATE_STOPPED" : `JOB_STATE_STOPPED` indicates that the job has
+ /// not
+ /// yet started to run.
+ /// - "JOB_STATE_RUNNING" : `JOB_STATE_RUNNING` indicates that the job is
+ /// currently running.
+ /// - "JOB_STATE_DONE" : `JOB_STATE_DONE` indicates that the job has
+ /// successfully completed.
+ /// This is a terminal job state. This state may be set by the Cloud
+ /// Dataflow
+ /// service, as a transition from `JOB_STATE_RUNNING`. It may also be set via
+ /// a
+ /// Cloud Dataflow `UpdateJob` call, if the job has not yet reached a
+ /// terminal
+ /// state.
+ /// - "JOB_STATE_FAILED" : `JOB_STATE_FAILED` indicates that the job has
+ /// failed. This is a
+ /// terminal job state. This state may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_CANCELLED" : `JOB_STATE_CANCELLED` indicates that the job
+ /// has been explicitly
+ /// cancelled. This is a terminal job state. This state may only be
+ /// set via a Cloud Dataflow `UpdateJob` call, and only if the job has not
+ /// yet reached another terminal state.
+ /// - "JOB_STATE_UPDATED" : `JOB_STATE_UPDATED` indicates that the job was
+ /// successfully updated,
+ /// meaning that this job was stopped and another job was started, inheriting
+ /// state from this one. This is a terminal job state. This state may only be
+ /// set by the Cloud Dataflow service, and only as a transition from
+ /// `JOB_STATE_RUNNING`.
+ /// - "JOB_STATE_DRAINING" : `JOB_STATE_DRAINING` indicates that the job is
+ /// in the process of draining.
+ /// A draining job has stopped pulling from its input sources and is
+ /// processing
+ /// any data that remains in-flight. This state may be set via a Cloud
+ /// Dataflow
+ /// `UpdateJob` call, but only as a transition from `JOB_STATE_RUNNING`. Jobs
+ /// that are draining may only transition to `JOB_STATE_DRAINED`,
+ /// `JOB_STATE_CANCELLED`, or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_DRAINED" : `JOB_STATE_DRAINED` indicates that the job has
+ /// been drained.
+ /// A drained job terminated by stopping pulling from its input sources and
+ /// processing any data that remained in-flight when draining was requested.
+ /// This state is a terminal state, may only be set by the Cloud Dataflow
+ /// service, and only as a transition from `JOB_STATE_DRAINING`.
+ /// - "JOB_STATE_PENDING" : 'JOB_STATE_PENDING' indicates that the job has
+ /// been created but is not yet
+ /// running. Jobs that are pending may only transition to
+ /// `JOB_STATE_RUNNING`,
+ /// or `JOB_STATE_FAILED`.
+ /// - "JOB_STATE_CANCELLING" : 'JOB_STATE_CANCELLING' indicates that the job
+ /// has been explicitly cancelled
+ /// and is in the process of stopping. Jobs that are cancelling may only
+ /// transition to 'JOB_STATE_CANCELLED' or 'JOB_STATE_FAILED'.
core.String requestedState;
- /**
- * This field may be mutated by the Cloud Dataflow service;
- * callers cannot mutate it.
- */
+
+ /// This field may be mutated by the Cloud Dataflow service;
+ /// callers cannot mutate it.
core.List<ExecutionStageState> stageStates;
- /** The top-level steps that constitute the entire job. */
+
+ /// The top-level steps that constitute the entire job.
core.List<Step> steps;
- /**
- * A set of files the system should be aware of that are used
- * for temporary storage. These temporary files will be
- * removed on job completion.
- * No duplicates are allowed.
- * No file patterns are supported.
- *
- * The supported files are:
- *
- * Google Cloud Storage:
- *
- * storage.googleapis.com/{bucket}/{object}
- * bucket.storage.googleapis.com/{object}
- */
+
+ /// A set of files the system should be aware of that are used
+ /// for temporary storage. These temporary files will be
+ /// removed on job completion.
+ /// No duplicates are allowed.
+ /// No file patterns are supported.
+ ///
+ /// The supported files are:
+ ///
+ /// Google Cloud Storage:
+ ///
+ /// storage.googleapis.com/{bucket}/{object}
+ /// bucket.storage.googleapis.com/{object}
core.List<core.String> tempFiles;
- /**
- * The map of transform name prefixes of the job to be replaced to the
- * corresponding name prefixes of the new job.
- */
+
+ /// The map of transform name prefixes of the job to be replaced to the
+ /// corresponding name prefixes of the new job.
core.Map<core.String, core.String> transformNameMapping;
- /**
- * The type of Cloud Dataflow job.
- * Possible string values are:
- * - "JOB_TYPE_UNKNOWN" : The type of the job is unspecified, or unknown.
- * - "JOB_TYPE_BATCH" : A batch job with a well-defined end point: data is
- * read, data is
- * processed, data is written, and the job is done.
- * - "JOB_TYPE_STREAMING" : A continuously streaming job with no end: data is
- * read,
- * processed, and written continuously.
- */
+
+ /// The type of Cloud Dataflow job.
+ /// Possible string values are:
+ /// - "JOB_TYPE_UNKNOWN" : The type of the job is unspecified, or unknown.
+ /// - "JOB_TYPE_BATCH" : A batch job with a well-defined end point: data is
+ /// read, data is
+ /// processed, data is written, and the job is done.
+ /// - "JOB_TYPE_STREAMING" : A continuously streaming job with no end: data
+ /// is read,
+ /// processed, and written continuously.
core.String type;
Job();
@@ -4173,7 +4350,8 @@ class Job {
name = _json["name"];
}
if (_json.containsKey("pipelineDescription")) {
- pipelineDescription = new PipelineDescription.fromJson(_json["pipelineDescription"]);
+ pipelineDescription =
+ new PipelineDescription.fromJson(_json["pipelineDescription"]);
}
if (_json.containsKey("projectId")) {
projectId = _json["projectId"];
@@ -4188,7 +4366,9 @@ class Job {
requestedState = _json["requestedState"];
}
if (_json.containsKey("stageStates")) {
- stageStates = _json["stageStates"].map((value) => new ExecutionStageState.fromJson(value)).toList();
+ stageStates = _json["stageStates"]
+ .map((value) => new ExecutionStageState.fromJson(value))
+ .toList();
}
if (_json.containsKey("steps")) {
steps = _json["steps"].map((value) => new Step.fromJson(value)).toList();
@@ -4205,7 +4385,8 @@ class Job {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (clientRequestId != null) {
_json["clientRequestId"] = clientRequestId;
}
@@ -4252,7 +4433,8 @@ class Job {
_json["requestedState"] = requestedState;
}
if (stageStates != null) {
- _json["stageStates"] = stageStates.map((value) => (value).toJson()).toList();
+ _json["stageStates"] =
+ stageStates.map((value) => (value).toJson()).toList();
}
if (steps != null) {
_json["steps"] = steps.map((value) => (value).toJson()).toList();
@@ -4270,41 +4452,42 @@ class Job {
}
}
-/**
- * Additional information about how a Cloud Dataflow job will be executed that
- * isn't contained in the submitted job.
- */
+/// Additional information about how a Cloud Dataflow job will be executed that
+/// isn't contained in the submitted job.
class JobExecutionInfo {
- /** A mapping from each stage to the information about that stage. */
+ /// A mapping from each stage to the information about that stage.
core.Map<core.String, JobExecutionStageInfo> stages;
JobExecutionInfo();
JobExecutionInfo.fromJson(core.Map _json) {
if (_json.containsKey("stages")) {
- stages = commons.mapMap<core.Map<core.String, core.Object>, JobExecutionStageInfo>(_json["stages"], (core.Map<core.String, core.Object> item) => new JobExecutionStageInfo.fromJson(item));
+ stages = commons
+ .mapMap<core.Map<core.String, core.Object>, JobExecutionStageInfo>(
+ _json["stages"],
+ (core.Map<core.String, core.Object> item) =>
+ new JobExecutionStageInfo.fromJson(item));
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (stages != null) {
- _json["stages"] = commons.mapMap<JobExecutionStageInfo, core.Map<core.String, core.Object>>(stages, (JobExecutionStageInfo item) => (item).toJson());
+ _json["stages"] = commons
+ .mapMap<JobExecutionStageInfo, core.Map<core.String, core.Object>>(
+ stages, (JobExecutionStageInfo item) => (item).toJson());
}
return _json;
}
}
-/**
- * Contains information about how a particular
- * google.dataflow.v1beta3.Step will be executed.
- */
+/// Contains information about how a particular
+/// google.dataflow.v1beta3.Step will be executed.
class JobExecutionStageInfo {
- /**
- * The steps associated with the execution stage.
- * Note that stages may have several steps, and that a given step
- * might be run by more than one stage.
- */
+ /// The steps associated with the execution stage.
+ /// Note that stages may have several steps, and that a given step
+ /// might be run by more than one stage.
core.List<core.String> stepName;
JobExecutionStageInfo();
@@ -4316,7 +4499,8 @@ class JobExecutionStageInfo {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (stepName != null) {
_json["stepName"] = stepName;
}
@@ -4324,48 +4508,49 @@ class JobExecutionStageInfo {
}
}
-/** A particular message pertaining to a Dataflow job. */
+/// A particular message pertaining to a Dataflow job.
class JobMessage {
- /** Deprecated. */
+ /// Deprecated.
core.String id;
- /**
- * Importance level of the message.
- * Possible string values are:
- * - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : The message importance isn't
- * specified, or is unknown.
- * - "JOB_MESSAGE_DEBUG" : The message is at the 'debug' level: typically only
- * useful for
- * software engineers working on the code the job is running.
- * Typically, Dataflow pipeline runners do not display log messages
- * at this level by default.
- * - "JOB_MESSAGE_DETAILED" : The message is at the 'detailed' level: somewhat
- * verbose, but
- * potentially useful to users. Typically, Dataflow pipeline
- * runners do not display log messages at this level by default.
- * These messages are displayed by default in the Dataflow
- * monitoring UI.
- * - "JOB_MESSAGE_BASIC" : The message is at the 'basic' level: useful for
- * keeping
- * track of the execution of a Dataflow pipeline. Typically,
- * Dataflow pipeline runners display log messages at this level by
- * default, and these messages are displayed by default in the
- * Dataflow monitoring UI.
- * - "JOB_MESSAGE_WARNING" : The message is at the 'warning' level: indicating
- * a condition
- * pertaining to a job which may require human intervention.
- * Typically, Dataflow pipeline runners display log messages at this
- * level by default, and these messages are displayed by default in
- * the Dataflow monitoring UI.
- * - "JOB_MESSAGE_ERROR" : The message is at the 'error' level: indicating a
- * condition
- * preventing a job from succeeding. Typically, Dataflow pipeline
- * runners display log messages at this level by default, and these
- * messages are displayed by default in the Dataflow monitoring UI.
- */
+
+ /// Importance level of the message.
+ /// Possible string values are:
+ /// - "JOB_MESSAGE_IMPORTANCE_UNKNOWN" : The message importance isn't
+ /// specified, or is unknown.
+ /// - "JOB_MESSAGE_DEBUG" : The message is at the 'debug' level: typically
+ /// only useful for
+ /// software engineers working on the code the job is running.
+ /// Typically, Dataflow pipeline runners do not display log messages
+ /// at this level by default.
+ /// - "JOB_MESSAGE_DETAILED" : The message is at the 'detailed' level:
+ /// somewhat verbose, but
+ /// potentially useful to users. Typically, Dataflow pipeline
+ /// runners do not display log messages at this level by default.
+ /// These messages are displayed by default in the Dataflow
+ /// monitoring UI.
+ /// - "JOB_MESSAGE_BASIC" : The message is at the 'basic' level: useful for
+ /// keeping
+ /// track of the execution of a Dataflow pipeline. Typically,
+ /// Dataflow pipeline runners display log messages at this level by
+ /// default, and these messages are displayed by default in the
+ /// Dataflow monitoring UI.
+ /// - "JOB_MESSAGE_WARNING" : The message is at the 'warning' level:
+ /// indicating a condition
+ /// pertaining to a job which may require human intervention.
+ /// Typically, Dataflow pipeline runners display log messages at this
+ /// level by default, and these messages are displayed by default in
+ /// the Dataflow monitoring UI.
+ /// - "JOB_MESSAGE_ERROR" : The message is at the 'error' level: indicating a
+ /// condition
+ /// preventing a job from succeeding. Typically, Dataflow pipeline
+ /// runners display log messages at this level by default, and these
+ /// messages are displayed by default in the Dataflow monitoring UI.
core.String messageImportance;
- /** The text of the message. */
+
+ /// The text of the message.
core.String messageText;
- /** The timestamp of the message. */
+
+ /// The timestamp of the message.
core.String time;
JobMessage();
@@ -4386,7 +4571,8 @@ class JobMessage {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (id != null) {
_json["id"] = id;
}
@@ -4403,19 +4589,18 @@ class JobMessage {
}
}
-/**
- * JobMetrics contains a collection of metrics descibing the detailed progress
- * of a Dataflow job. Metrics correspond to user-defined and system-defined
- * metrics in the job.
- *
- * This resource captures only the most recent values of each metric;
- * time-series data can be queried for them (under the same metric names)
- * from Cloud Monitoring.
- */
+/// JobMetrics contains a collection of metrics descibing the detailed progress
+/// of a Dataflow job. Metrics correspond to user-defined and system-defined
+/// metrics in the job.
+///
+/// This resource captures only the most recent values of each metric;
+/// time-series data can be queried for them (under the same metric names)
+/// from Cloud Monitoring.
class JobMetrics {
- /** Timestamp as of which metric values are current. */
+ /// Timestamp as of which metric values are current.
core.String metricTime;
- /** All metrics for this job. */
+
+ /// All metrics for this job.
core.List<MetricUpdate> metrics;
JobMetrics();
@@ -4425,12 +4610,15 @@ class JobMetrics {
metricTime = _json["metricTime"];
}
if (_json.containsKey("metrics")) {
- metrics = _json["metrics"].map((value) => new MetricUpdate.fromJson(value)).toList();
+ metrics = _json["metrics"]
+ .map((value) => new MetricUpdate.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (metricTime != null) {
_json["metricTime"] = metricTime;
}
@@ -4441,23 +4629,21 @@ class JobMetrics {
}
}
-/**
- * Data disk assignment information for a specific key-range of a sharded
- * computation.
- * Currently we only support UTF-8 character splits to simplify encoding into
- * JSON.
- */
+/// Data disk assignment information for a specific key-range of a sharded
+/// computation.
+/// Currently we only support UTF-8 character splits to simplify encoding into
+/// JSON.
class KeyRangeDataDiskAssignment {
- /**
- * The name of the data disk where data for this range is stored.
- * This name is local to the Google Cloud Platform project and uniquely
- * identifies the disk within that project, for example
- * "myproject-1014-104817-4c2-harness-0-disk-1".
- */
+ /// The name of the data disk where data for this range is stored.
+ /// This name is local to the Google Cloud Platform project and uniquely
+ /// identifies the disk within that project, for example
+ /// "myproject-1014-104817-4c2-harness-0-disk-1".
core.String dataDisk;
- /** The end (exclusive) of the key range. */
+
+ /// The end (exclusive) of the key range.
core.String end;
- /** The start (inclusive) of the key range. */
+
+ /// The start (inclusive) of the key range.
core.String start;
KeyRangeDataDiskAssignment();
@@ -4475,7 +4661,8 @@ class KeyRangeDataDiskAssignment {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dataDisk != null) {
_json["dataDisk"] = dataDisk;
}
@@ -4489,32 +4676,28 @@ class KeyRangeDataDiskAssignment {
}
}
-/**
- * Location information for a specific key-range of a sharded computation.
- * Currently we only support UTF-8 character splits to simplify encoding into
- * JSON.
- */
+/// Location information for a specific key-range of a sharded computation.
+/// Currently we only support UTF-8 character splits to simplify encoding into
+/// JSON.
class KeyRangeLocation {
- /**
- * The name of the data disk where data for this range is stored.
- * This name is local to the Google Cloud Platform project and uniquely
- * identifies the disk within that project, for example
- * "myproject-1014-104817-4c2-harness-0-disk-1".
- */
+ /// The name of the data disk where data for this range is stored.
+ /// This name is local to the Google Cloud Platform project and uniquely
+ /// identifies the disk within that project, for example
+ /// "myproject-1014-104817-4c2-harness-0-disk-1".
core.String dataDisk;
- /**
- * The physical location of this range assignment to be used for
- * streaming computation cross-worker message delivery.
- */
+
+ /// The physical location of this range assignment to be used for
+ /// streaming computation cross-worker message delivery.
core.String deliveryEndpoint;
- /**
- * DEPRECATED. The location of the persistent state for this range, as a
- * persistent directory in the worker local filesystem.
- */
+
+ /// DEPRECATED. The location of the persistent state for this range, as a
+ /// persistent directory in the worker local filesystem.
core.String deprecatedPersistentDirectory;
- /** The end (exclusive) of the key range. */
+
+ /// The end (exclusive) of the key range.
core.String end;
- /** The start (inclusive) of the key range. */
+
+ /// The start (inclusive) of the key range.
core.String start;
KeyRangeLocation();
@@ -4538,7 +4721,8 @@ class KeyRangeLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dataDisk != null) {
_json["dataDisk"] = dataDisk;
}
@@ -4558,13 +4742,15 @@ class KeyRangeLocation {
}
}
-/** Parameters to provide to the template being launched. */
+/// Parameters to provide to the template being launched.
class LaunchTemplateParameters {
- /** The runtime environment for the job. */
+ /// The runtime environment for the job.
RuntimeEnvironment environment;
- /** Required. The job name to use for the created job. */
+
+ /// Required. The job name to use for the created job.
core.String jobName;
- /** The runtime parameters to pass to the job. */
+
+ /// The runtime parameters to pass to the job.
core.Map<core.String, core.String> parameters;
LaunchTemplateParameters();
@@ -4582,7 +4768,8 @@ class LaunchTemplateParameters {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (environment != null) {
_json["environment"] = (environment).toJson();
}
@@ -4596,12 +4783,10 @@ class LaunchTemplateParameters {
}
}
-/** Response to the request to launch a template. */
+/// Response to the request to launch a template.
class LaunchTemplateResponse {
- /**
- * The job that was launched, if the request was not a dry run and
- * the job was successfully launched.
- */
+ /// The job that was launched, if the request was not a dry run and
+ /// the job was successfully launched.
Job job;
LaunchTemplateResponse();
@@ -4613,7 +4798,8 @@ class LaunchTemplateResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (job != null) {
_json["job"] = (job).toJson();
}
@@ -4621,25 +4807,26 @@ class LaunchTemplateResponse {
}
}
-/** Request to lease WorkItems. */
+/// Request to lease WorkItems.
class LeaseWorkItemRequest {
- /** The current timestamp at the worker. */
+ /// The current timestamp at the worker.
core.String currentWorkerTime;
- /** The location which contains the WorkItem's job. */
+
+ /// The location which contains the WorkItem's job.
core.String location;
- /** The initial lease period. */
+
+ /// The initial lease period.
core.String requestedLeaseDuration;
- /** Filter for WorkItem type. */
+
+ /// Filter for WorkItem type.
core.List<core.String> workItemTypes;
- /**
- * Worker capabilities. WorkItems might be limited to workers with specific
- * capabilities.
- */
+
+ /// Worker capabilities. WorkItems might be limited to workers with specific
+ /// capabilities.
core.List<core.String> workerCapabilities;
- /**
- * Identifies the worker leasing work -- typically the ID of the
- * virtual machine running the worker.
- */
+
+ /// Identifies the worker leasing work -- typically the ID of the
+ /// virtual machine running the worker.
core.String workerId;
LeaseWorkItemRequest();
@@ -4666,7 +4853,8 @@ class LeaseWorkItemRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (currentWorkerTime != null) {
_json["currentWorkerTime"] = currentWorkerTime;
}
@@ -4689,21 +4877,24 @@ class LeaseWorkItemRequest {
}
}
-/** Response to a request to lease WorkItems. */
+/// Response to a request to lease WorkItems.
class LeaseWorkItemResponse {
- /** A list of the leased WorkItems. */
+ /// A list of the leased WorkItems.
core.List<WorkItem> workItems;
LeaseWorkItemResponse();
LeaseWorkItemResponse.fromJson(core.Map _json) {
if (_json.containsKey("workItems")) {
- workItems = _json["workItems"].map((value) => new WorkItem.fromJson(value)).toList();
+ workItems = _json["workItems"]
+ .map((value) => new WorkItem.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (workItems != null) {
_json["workItems"] = workItems.map((value) => (value).toJson()).toList();
}
@@ -4711,23 +4902,29 @@ class LeaseWorkItemResponse {
}
}
-/** Response to a request to list job messages. */
+/// Response to a request to list job messages.
class ListJobMessagesResponse {
- /** Autoscaling events in ascending timestamp order. */
+ /// Autoscaling events in ascending timestamp order.
core.List<AutoscalingEvent> autoscalingEvents;
- /** Messages in ascending timestamp order. */
+
+ /// Messages in ascending timestamp order.
core.List<JobMessage> jobMessages;
- /** The token to obtain the next page of results if there are more. */
+
+ /// The token to obtain the next page of results if there are more.
core.String nextPageToken;
ListJobMessagesResponse();
ListJobMessagesResponse.fromJson(core.Map _json) {
if (_json.containsKey("autoscalingEvents")) {
- autoscalingEvents = _json["autoscalingEvents"].map((value) => new AutoscalingEvent.fromJson(value)).toList();
+ autoscalingEvents = _json["autoscalingEvents"]
+ .map((value) => new AutoscalingEvent.fromJson(value))
+ .toList();
}
if (_json.containsKey("jobMessages")) {
- jobMessages = _json["jobMessages"].map((value) => new JobMessage.fromJson(value)).toList();
+ jobMessages = _json["jobMessages"]
+ .map((value) => new JobMessage.fromJson(value))
+ .toList();
}
if (_json.containsKey("nextPageToken")) {
nextPageToken = _json["nextPageToken"];
@@ -4735,12 +4932,15 @@ class ListJobMessagesResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (autoscalingEvents != null) {
- _json["autoscalingEvents"] = autoscalingEvents.map((value) => (value).toJson()).toList();
+ _json["autoscalingEvents"] =
+ autoscalingEvents.map((value) => (value).toJson()).toList();
}
if (jobMessages != null) {
- _json["jobMessages"] = jobMessages.map((value) => (value).toJson()).toList();
+ _json["jobMessages"] =
+ jobMessages.map((value) => (value).toJson()).toList();
}
if (nextPageToken != null) {
_json["nextPageToken"] = nextPageToken;
@@ -4749,23 +4949,25 @@ class ListJobMessagesResponse {
}
}
-/**
- * Response to a request to list Cloud Dataflow jobs. This may be a partial
- * response, depending on the page size in the ListJobsRequest.
- */
+/// Response to a request to list Cloud Dataflow jobs. This may be a partial
+/// response, depending on the page size in the ListJobsRequest.
class ListJobsResponse {
- /** Zero or more messages describing locations that failed to respond. */
+ /// Zero or more messages describing locations that failed to respond.
core.List<FailedLocation> failedLocation;
- /** A subset of the requested job information. */
+
+ /// A subset of the requested job information.
core.List<Job> jobs;
- /** Set if there may be more results than fit in this response. */
+
+ /// Set if there may be more results than fit in this response.
core.String nextPageToken;
ListJobsResponse();
ListJobsResponse.fromJson(core.Map _json) {
if (_json.containsKey("failedLocation")) {
- failedLocation = _json["failedLocation"].map((value) => new FailedLocation.fromJson(value)).toList();
+ failedLocation = _json["failedLocation"]
+ .map((value) => new FailedLocation.fromJson(value))
+ .toList();
}
if (_json.containsKey("jobs")) {
jobs = _json["jobs"].map((value) => new Job.fromJson(value)).toList();
@@ -4776,9 +4978,11 @@ class ListJobsResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (failedLocation != null) {
- _json["failedLocation"] = failedLocation.map((value) => (value).toJson()).toList();
+ _json["failedLocation"] =
+ failedLocation.map((value) => (value).toJson()).toList();
}
if (jobs != null) {
_json["jobs"] = jobs.map((value) => (value).toJson()).toList();
@@ -4790,20 +4994,19 @@ class ListJobsResponse {
}
}
-/** Bucket of values for Distribution's logarithmic histogram. */
+/// Bucket of values for Distribution's logarithmic histogram.
class LogBucket {
- /** Number of values in this bucket. */
+ /// Number of values in this bucket.
core.String count;
- /**
- * floor(log2(value)); defined to be zero for nonpositive values.
- * log(-1) = 0
- * log(0) = 0
- * log(1) = 0
- * log(2) = 1
- * log(3) = 1
- * log(4) = 2
- * log(5) = 2
- */
+
+ /// floor(log2(value)); defined to be zero for nonpositive values.
+ /// log(-1) = 0
+ /// log(0) = 0
+ /// log(1) = 0
+ /// log(2) = 1
+ /// log(3) = 1
+ /// log(4) = 2
+ /// log(5) = 2
core.int log;
LogBucket();
@@ -4818,7 +5021,8 @@ class LogBucket {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (count != null) {
_json["count"] = count;
}
@@ -4829,33 +5033,31 @@ class LogBucket {
}
}
-/**
- * MapTask consists of an ordered set of instructions, each of which
- * describes one particular low-level operation for the worker to
- * perform in order to accomplish the MapTask's WorkItem.
- *
- * Each instruction must appear in the list before any instructions which
- * depends on its output.
- */
+/// MapTask consists of an ordered set of instructions, each of which
+/// describes one particular low-level operation for the worker to
+/// perform in order to accomplish the MapTask's WorkItem.
+///
+/// Each instruction must appear in the list before any instructions which
+/// depends on its output.
class MapTask {
- /** The instructions in the MapTask. */
+ /// The instructions in the MapTask.
core.List<ParallelInstruction> instructions;
- /**
- * System-defined name of the stage containing this MapTask.
- * Unique across the workflow.
- */
+
+ /// System-defined name of the stage containing this MapTask.
+ /// Unique across the workflow.
core.String stageName;
- /**
- * System-defined name of this MapTask.
- * Unique across the workflow.
- */
+
+ /// System-defined name of this MapTask.
+ /// Unique across the workflow.
core.String systemName;
MapTask();
MapTask.fromJson(core.Map _json) {
if (_json.containsKey("instructions")) {
- instructions = _json["instructions"].map((value) => new ParallelInstruction.fromJson(value)).toList();
+ instructions = _json["instructions"]
+ .map((value) => new ParallelInstruction.fromJson(value))
+ .toList();
}
if (_json.containsKey("stageName")) {
stageName = _json["stageName"];
@@ -4866,9 +5068,11 @@ class MapTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (instructions != null) {
- _json["instructions"] = instructions.map((value) => (value).toJson()).toList();
+ _json["instructions"] =
+ instructions.map((value) => (value).toJson()).toList();
}
if (stageName != null) {
_json["stageName"] = stageName;
@@ -4880,17 +5084,14 @@ class MapTask {
}
}
-/**
- * The metric short id is returned to the user alongside an offset into
- * ReportWorkItemStatusRequest
- */
+/// The metric short id is returned to the user alongside an offset into
+/// ReportWorkItemStatusRequest
class MetricShortId {
- /**
- * The index of the corresponding metric in
- * the ReportWorkItemStatusRequest. Required.
- */
+ /// The index of the corresponding metric in
+ /// the ReportWorkItemStatusRequest. Required.
core.int metricIndex;
- /** The service-generated short identifier for the metric. */
+
+ /// The service-generated short identifier for the metric.
core.String shortId;
MetricShortId();
@@ -4905,7 +5106,8 @@ class MetricShortId {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (metricIndex != null) {
_json["metricIndex"] = metricIndex;
}
@@ -4916,26 +5118,22 @@ class MetricShortId {
}
}
-/**
- * Identifies a metric, by describing the source which generated the
- * metric.
- */
+/// Identifies a metric, by describing the source which generated the
+/// metric.
class MetricStructuredName {
- /**
- * Zero or more labeled fields which identify the part of the job this
- * metric is associated with, such as the name of a step or collection.
- *
- * For example, built-in counters associated with steps will have
- * context['step'] = <step-name>. Counters associated with PCollections
- * in the SDK will have context['pcollection'] = <pcollection-name>.
- */
+ /// Zero or more labeled fields which identify the part of the job this
+ /// metric is associated with, such as the name of a step or collection.
+ ///
+ /// For example, built-in counters associated with steps will have
+ /// context['step'] = <step-name>. Counters associated with PCollections
+ /// in the SDK will have context['pcollection'] = <pcollection-name>.
core.Map<core.String, core.String> context;
- /** Worker-defined metric name. */
+
+ /// Worker-defined metric name.
core.String name;
- /**
- * Origin (namespace) of metric name. May be blank for user-define metrics;
- * will be "dataflow" for metrics defined by the Dataflow service or SDK.
- */
+
+ /// Origin (namespace) of metric name. May be blank for user-define metrics;
+ /// will be "dataflow" for metrics defined by the Dataflow service or SDK.
core.String origin;
MetricStructuredName();
@@ -4953,7 +5151,8 @@ class MetricStructuredName {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (context != null) {
_json["context"] = context;
}
@@ -4967,84 +5166,76 @@ class MetricStructuredName {
}
}
-/** Describes the state of a metric. */
+/// Describes the state of a metric.
class MetricUpdate {
- /**
- * True if this metric is reported as the total cumulative aggregate
- * value accumulated since the worker started working on this WorkItem.
- * By default this is false, indicating that this metric is reported
- * as a delta that is not associated with any WorkItem.
- */
+ /// True if this metric is reported as the total cumulative aggregate
+ /// value accumulated since the worker started working on this WorkItem.
+ /// By default this is false, indicating that this metric is reported
+ /// as a delta that is not associated with any WorkItem.
core.bool cumulative;
- /**
- * A struct value describing properties of a distribution of numeric values.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A struct value describing properties of a distribution of numeric values.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object distribution;
- /**
- * Worker-computed aggregate value for internal use by the Dataflow
- * service.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Worker-computed aggregate value for internal use by the Dataflow
+ /// service.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object internal;
- /**
- * Metric aggregation kind. The possible metric aggregation kinds are
- * "Sum", "Max", "Min", "Mean", "Set", "And", "Or", and "Distribution".
- * The specified aggregation kind is case-insensitive.
- *
- * If omitted, this is not an aggregated value but instead
- * a single metric sample value.
- */
+
+ /// Metric aggregation kind. The possible metric aggregation kinds are
+ /// "Sum", "Max", "Min", "Mean", "Set", "And", "Or", and "Distribution".
+ /// The specified aggregation kind is case-insensitive.
+ ///
+ /// If omitted, this is not an aggregated value but instead
+ /// a single metric sample value.
core.String kind;
- /**
- * Worker-computed aggregate value for the "Mean" aggregation kind.
- * This holds the count of the aggregated values and is used in combination
- * with mean_sum above to obtain the actual mean aggregate value.
- * The only possible value type is Long.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Worker-computed aggregate value for the "Mean" aggregation kind.
+ /// This holds the count of the aggregated values and is used in combination
+ /// with mean_sum above to obtain the actual mean aggregate value.
+ /// The only possible value type is Long.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object meanCount;
- /**
- * Worker-computed aggregate value for the "Mean" aggregation kind.
- * This holds the sum of the aggregated values and is used in combination
- * with mean_count below to obtain the actual mean aggregate value.
- * The only possible value types are Long and Double.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Worker-computed aggregate value for the "Mean" aggregation kind.
+ /// This holds the sum of the aggregated values and is used in combination
+ /// with mean_count below to obtain the actual mean aggregate value.
+ /// The only possible value types are Long and Double.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object meanSum;
- /** Name of the metric. */
+
+ /// Name of the metric.
MetricStructuredName name;
- /**
- * Worker-computed aggregate value for aggregation kinds "Sum", "Max", "Min",
- * "And", and "Or". The possible value types are Long, Double, and Boolean.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Worker-computed aggregate value for aggregation kinds "Sum", "Max",
+ /// "Min",
+ /// "And", and "Or". The possible value types are Long, Double, and Boolean.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object scalar;
- /**
- * Worker-computed aggregate value for the "Set" aggregation kind. The only
- * possible value type is a list of Values whose type can be Long, Double,
- * or String, according to the metric's type. All Values in the list must
- * be of the same type.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Worker-computed aggregate value for the "Set" aggregation kind. The only
+ /// possible value type is a list of Values whose type can be Long, Double,
+ /// or String, according to the metric's type. All Values in the list must
+ /// be of the same type.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object set;
- /**
- * Timestamp associated with the metric value. Optional when workers are
- * reporting work progress; it will be filled in responses from the
- * metrics API.
- */
+
+ /// Timestamp associated with the metric value. Optional when workers are
+ /// reporting work progress; it will be filled in responses from the
+ /// metrics API.
core.String updateTime;
MetricUpdate();
@@ -5083,7 +5274,8 @@ class MetricUpdate {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (cumulative != null) {
_json["cumulative"] = cumulative;
}
@@ -5118,14 +5310,12 @@ class MetricUpdate {
}
}
-/** Describes mounted data disk. */
+/// Describes mounted data disk.
class MountedDataDisk {
- /**
- * The name of the data disk.
- * This name is local to the Google Cloud Platform project and uniquely
- * identifies the disk within that project, for example
- * "myproject-1014-104817-4c2-harness-0-disk-1".
- */
+ /// The name of the data disk.
+ /// This name is local to the Google Cloud Platform project and uniquely
+ /// identifies the disk within that project, for example
+ /// "myproject-1014-104817-4c2-harness-0-disk-1".
core.String dataDisk;
MountedDataDisk();
@@ -5137,7 +5327,8 @@ class MountedDataDisk {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dataDisk != null) {
_json["dataDisk"] = dataDisk;
}
@@ -5145,12 +5336,10 @@ class MountedDataDisk {
}
}
-/** Information about an output of a multi-output DoFn. */
+/// Information about an output of a multi-output DoFn.
class MultiOutputInfo {
- /**
- * The id of the tag the user code will emit to this output by; this
- * should correspond to the tag of some SideInputInfo.
- */
+ /// The id of the tag the user code will emit to this output by; this
+ /// should correspond to the tag of some SideInputInfo.
core.String tag;
MultiOutputInfo();
@@ -5162,7 +5351,8 @@ class MultiOutputInfo {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (tag != null) {
_json["tag"] = tag;
}
@@ -5170,26 +5360,25 @@ class MultiOutputInfo {
}
}
-/** Basic metadata about a counter. */
+/// Basic metadata about a counter.
class NameAndKind {
- /**
- * Counter aggregation kind.
- * Possible string values are:
- * - "INVALID" : Counter aggregation kind was not set.
- * - "SUM" : Aggregated value is the sum of all contributed values.
- * - "MAX" : Aggregated value is the max of all contributed values.
- * - "MIN" : Aggregated value is the min of all contributed values.
- * - "MEAN" : Aggregated value is the mean of all contributed values.
- * - "OR" : Aggregated value represents the logical 'or' of all contributed
- * values.
- * - "AND" : Aggregated value represents the logical 'and' of all contributed
- * values.
- * - "SET" : Aggregated value is a set of unique contributed values.
- * - "DISTRIBUTION" : Aggregated value captures statistics about a
- * distribution.
- */
+ /// Counter aggregation kind.
+ /// Possible string values are:
+ /// - "INVALID" : Counter aggregation kind was not set.
+ /// - "SUM" : Aggregated value is the sum of all contributed values.
+ /// - "MAX" : Aggregated value is the max of all contributed values.
+ /// - "MIN" : Aggregated value is the min of all contributed values.
+ /// - "MEAN" : Aggregated value is the mean of all contributed values.
+ /// - "OR" : Aggregated value represents the logical 'or' of all contributed
+ /// values.
+ /// - "AND" : Aggregated value represents the logical 'and' of all
+ /// contributed values.
+ /// - "SET" : Aggregated value is a set of unique contributed values.
+ /// - "DISTRIBUTION" : Aggregated value captures statistics about a
+ /// distribution.
core.String kind;
- /** Name of the counter. */
+
+ /// Name of the counter.
core.String name;
NameAndKind();
@@ -5204,7 +5393,8 @@ class NameAndKind {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (kind != null) {
_json["kind"] = kind;
}
@@ -5215,28 +5405,25 @@ class NameAndKind {
}
}
-/**
- * The packages that must be installed in order for a worker to run the
- * steps of the Cloud Dataflow job that will be assigned to its worker
- * pool.
- *
- * This is the mechanism by which the Cloud Dataflow SDK causes code to
- * be loaded onto the workers. For example, the Cloud Dataflow Java SDK
- * might use this to install jars containing the user's code and all of the
- * various dependencies (libraries, data files, etc.) required in order
- * for that code to run.
- */
+/// The packages that must be installed in order for a worker to run the
+/// steps of the Cloud Dataflow job that will be assigned to its worker
+/// pool.
+///
+/// This is the mechanism by which the Cloud Dataflow SDK causes code to
+/// be loaded onto the workers. For example, the Cloud Dataflow Java SDK
+/// might use this to install jars containing the user's code and all of the
+/// various dependencies (libraries, data files, etc.) required in order
+/// for that code to run.
class Package {
- /**
- * The resource to read the package from. The supported resource type is:
- *
- * Google Cloud Storage:
- *
- * storage.googleapis.com/{bucket}
- * bucket.storage.googleapis.com/
- */
+ /// The resource to read the package from. The supported resource type is:
+ ///
+ /// Google Cloud Storage:
+ ///
+ /// storage.googleapis.com/{bucket}
+ /// bucket.storage.googleapis.com/
core.String location;
- /** The name of the package. */
+
+ /// The name of the package.
core.String name;
Package();
@@ -5251,7 +5438,8 @@ class Package {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (location != null) {
_json["location"] = location;
}
@@ -5262,27 +5450,27 @@ class Package {
}
}
-/**
- * An instruction that does a ParDo operation.
- * Takes one main input and zero or more side inputs, and produces
- * zero or more outputs.
- * Runs user code.
- */
+/// An instruction that does a ParDo operation.
+/// Takes one main input and zero or more side inputs, and produces
+/// zero or more outputs.
+/// Runs user code.
class ParDoInstruction {
- /** The input. */
+ /// The input.
InstructionInput input;
- /** Information about each of the outputs, if user_fn is a MultiDoFn. */
+
+ /// Information about each of the outputs, if user_fn is a MultiDoFn.
core.List<MultiOutputInfo> multiOutputInfos;
- /** The number of outputs. */
+
+ /// The number of outputs.
core.int numOutputs;
- /** Zero or more side inputs. */
+
+ /// Zero or more side inputs.
core.List<SideInputInfo> sideInputs;
- /**
- * The user function to invoke.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The user function to invoke.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> userFn;
ParDoInstruction();
@@ -5292,13 +5480,17 @@ class ParDoInstruction {
input = new InstructionInput.fromJson(_json["input"]);
}
if (_json.containsKey("multiOutputInfos")) {
- multiOutputInfos = _json["multiOutputInfos"].map((value) => new MultiOutputInfo.fromJson(value)).toList();
+ multiOutputInfos = _json["multiOutputInfos"]
+ .map((value) => new MultiOutputInfo.fromJson(value))
+ .toList();
}
if (_json.containsKey("numOutputs")) {
numOutputs = _json["numOutputs"];
}
if (_json.containsKey("sideInputs")) {
- sideInputs = _json["sideInputs"].map((value) => new SideInputInfo.fromJson(value)).toList();
+ sideInputs = _json["sideInputs"]
+ .map((value) => new SideInputInfo.fromJson(value))
+ .toList();
}
if (_json.containsKey("userFn")) {
userFn = _json["userFn"];
@@ -5306,18 +5498,21 @@ class ParDoInstruction {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (input != null) {
_json["input"] = (input).toJson();
}
if (multiOutputInfos != null) {
- _json["multiOutputInfos"] = multiOutputInfos.map((value) => (value).toJson()).toList();
+ _json["multiOutputInfos"] =
+ multiOutputInfos.map((value) => (value).toJson()).toList();
}
if (numOutputs != null) {
_json["numOutputs"] = numOutputs;
}
if (sideInputs != null) {
- _json["sideInputs"] = sideInputs.map((value) => (value).toJson()).toList();
+ _json["sideInputs"] =
+ sideInputs.map((value) => (value).toJson()).toList();
}
if (userFn != null) {
_json["userFn"] = userFn;
@@ -5326,28 +5521,34 @@ class ParDoInstruction {
}
}
-/** Describes a particular operation comprising a MapTask. */
+/// Describes a particular operation comprising a MapTask.
class ParallelInstruction {
- /** Additional information for Flatten instructions. */
+ /// Additional information for Flatten instructions.
FlattenInstruction flatten;
- /** User-provided name of this operation. */
+
+ /// User-provided name of this operation.
core.String name;
- /** System-defined name for the operation in the original workflow graph. */
+
+ /// System-defined name for the operation in the original workflow graph.
core.String originalName;
- /** Describes the outputs of the instruction. */
+
+ /// Describes the outputs of the instruction.
core.List<InstructionOutput> outputs;
- /** Additional information for ParDo instructions. */
+
+ /// Additional information for ParDo instructions.
ParDoInstruction parDo;
- /** Additional information for PartialGroupByKey instructions. */
+
+ /// Additional information for PartialGroupByKey instructions.
PartialGroupByKeyInstruction partialGroupByKey;
- /** Additional information for Read instructions. */
+
+ /// Additional information for Read instructions.
ReadInstruction read;
- /**
- * System-defined name of this operation.
- * Unique across the workflow.
- */
+
+ /// System-defined name of this operation.
+ /// Unique across the workflow.
core.String systemName;
- /** Additional information for Write instructions. */
+
+ /// Additional information for Write instructions.
WriteInstruction write;
ParallelInstruction();
@@ -5363,13 +5564,16 @@ class ParallelInstruction {
originalName = _json["originalName"];
}
if (_json.containsKey("outputs")) {
- outputs = _json["outputs"].map((value) => new InstructionOutput.fromJson(value)).toList();
+ outputs = _json["outputs"]
+ .map((value) => new InstructionOutput.fromJson(value))
+ .toList();
}
if (_json.containsKey("parDo")) {
parDo = new ParDoInstruction.fromJson(_json["parDo"]);
}
if (_json.containsKey("partialGroupByKey")) {
- partialGroupByKey = new PartialGroupByKeyInstruction.fromJson(_json["partialGroupByKey"]);
+ partialGroupByKey =
+ new PartialGroupByKeyInstruction.fromJson(_json["partialGroupByKey"]);
}
if (_json.containsKey("read")) {
read = new ReadInstruction.fromJson(_json["read"]);
@@ -5383,7 +5587,8 @@ class ParallelInstruction {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (flatten != null) {
_json["flatten"] = (flatten).toJson();
}
@@ -5415,16 +5620,15 @@ class ParallelInstruction {
}
}
-/** Structured data associated with this message. */
+/// Structured data associated with this message.
class Parameter {
- /** Key or name for this parameter. */
+ /// Key or name for this parameter.
core.String key;
- /**
- * Value for this parameter.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Value for this parameter.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Object value;
Parameter();
@@ -5439,7 +5643,8 @@ class Parameter {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (key != null) {
_json["key"] = key;
}
@@ -5450,17 +5655,21 @@ class Parameter {
}
}
-/** Metadata for a specific parameter. */
+/// Metadata for a specific parameter.
class ParameterMetadata {
- /** Required. The help text to display for the parameter. */
+ /// Required. The help text to display for the parameter.
core.String helpText;
- /** Optional. Whether the parameter is optional. Defaults to false. */
+
+ /// Optional. Whether the parameter is optional. Defaults to false.
core.bool isOptional;
- /** Required. The label to display for the parameter. */
+
+ /// Required. The label to display for the parameter.
core.String label;
- /** Required. The name of the parameter. */
+
+ /// Required. The name of the parameter.
core.String name;
- /** Optional. Regexes that the parameter must match. */
+
+ /// Optional. Regexes that the parameter must match.
core.List<core.String> regexes;
ParameterMetadata();
@@ -5484,7 +5693,8 @@ class ParameterMetadata {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (helpText != null) {
_json["helpText"] = helpText;
}
@@ -5504,38 +5714,34 @@ class ParameterMetadata {
}
}
-/**
- * An instruction that does a partial group-by-key.
- * One input and one output.
- */
+/// An instruction that does a partial group-by-key.
+/// One input and one output.
class PartialGroupByKeyInstruction {
- /** Describes the input to the partial group-by-key instruction. */
+ /// Describes the input to the partial group-by-key instruction.
InstructionInput input;
- /**
- * The codec to use for interpreting an element in the input PTable.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The codec to use for interpreting an element in the input PTable.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> inputElementCodec;
- /**
- * If this instruction includes a combining function this is the name of the
- * intermediate store between the GBK and the CombineValues.
- */
+
+ /// If this instruction includes a combining function this is the name of the
+ /// intermediate store between the GBK and the CombineValues.
core.String originalCombineValuesInputStoreName;
- /**
- * If this instruction includes a combining function, this is the name of the
- * CombineValues instruction lifted into this instruction.
- */
+
+ /// If this instruction includes a combining function, this is the name of
+ /// the
+ /// CombineValues instruction lifted into this instruction.
core.String originalCombineValuesStepName;
- /** Zero or more side inputs. */
+
+ /// Zero or more side inputs.
core.List<SideInputInfo> sideInputs;
- /**
- * The value combining function to invoke.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The value combining function to invoke.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> valueCombiningFn;
PartialGroupByKeyInstruction();
@@ -5548,13 +5754,16 @@ class PartialGroupByKeyInstruction {
inputElementCodec = _json["inputElementCodec"];
}
if (_json.containsKey("originalCombineValuesInputStoreName")) {
- originalCombineValuesInputStoreName = _json["originalCombineValuesInputStoreName"];
+ originalCombineValuesInputStoreName =
+ _json["originalCombineValuesInputStoreName"];
}
if (_json.containsKey("originalCombineValuesStepName")) {
originalCombineValuesStepName = _json["originalCombineValuesStepName"];
}
if (_json.containsKey("sideInputs")) {
- sideInputs = _json["sideInputs"].map((value) => new SideInputInfo.fromJson(value)).toList();
+ sideInputs = _json["sideInputs"]
+ .map((value) => new SideInputInfo.fromJson(value))
+ .toList();
}
if (_json.containsKey("valueCombiningFn")) {
valueCombiningFn = _json["valueCombiningFn"];
@@ -5562,7 +5771,8 @@ class PartialGroupByKeyInstruction {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (input != null) {
_json["input"] = (input).toJson();
}
@@ -5570,13 +5780,15 @@ class PartialGroupByKeyInstruction {
_json["inputElementCodec"] = inputElementCodec;
}
if (originalCombineValuesInputStoreName != null) {
- _json["originalCombineValuesInputStoreName"] = originalCombineValuesInputStoreName;
+ _json["originalCombineValuesInputStoreName"] =
+ originalCombineValuesInputStoreName;
}
if (originalCombineValuesStepName != null) {
_json["originalCombineValuesStepName"] = originalCombineValuesStepName;
}
if (sideInputs != null) {
- _json["sideInputs"] = sideInputs.map((value) => (value).toJson()).toList();
+ _json["sideInputs"] =
+ sideInputs.map((value) => (value).toJson()).toList();
}
if (valueCombiningFn != null) {
_json["valueCombiningFn"] = valueCombiningFn;
@@ -5585,73 +5797,82 @@ class PartialGroupByKeyInstruction {
}
}
-/**
- * A descriptive representation of submitted pipeline as well as the executed
- * form. This data is provided by the Dataflow service for ease of visualizing
- * the pipeline and interpretting Dataflow provided metrics.
- */
+/// A descriptive representation of submitted pipeline as well as the executed
+/// form. This data is provided by the Dataflow service for ease of
+/// visualizing
+/// the pipeline and interpretting Dataflow provided metrics.
class PipelineDescription {
- /** Pipeline level display data. */
+ /// Pipeline level display data.
core.List<DisplayData> displayData;
- /** Description of each stage of execution of the pipeline. */
+
+ /// Description of each stage of execution of the pipeline.
core.List<ExecutionStageSummary> executionPipelineStage;
- /**
- * Description of each transform in the pipeline and collections between them.
- */
+
+ /// Description of each transform in the pipeline and collections between
+ /// them.
core.List<TransformSummary> originalPipelineTransform;
PipelineDescription();
PipelineDescription.fromJson(core.Map _json) {
if (_json.containsKey("displayData")) {
- displayData = _json["displayData"].map((value) => new DisplayData.fromJson(value)).toList();
+ displayData = _json["displayData"]
+ .map((value) => new DisplayData.fromJson(value))
+ .toList();
}
if (_json.containsKey("executionPipelineStage")) {
- executionPipelineStage = _json["executionPipelineStage"].map((value) => new ExecutionStageSummary.fromJson(value)).toList();
+ executionPipelineStage = _json["executionPipelineStage"]
+ .map((value) => new ExecutionStageSummary.fromJson(value))
+ .toList();
}
if (_json.containsKey("originalPipelineTransform")) {
- originalPipelineTransform = _json["originalPipelineTransform"].map((value) => new TransformSummary.fromJson(value)).toList();
+ originalPipelineTransform = _json["originalPipelineTransform"]
+ .map((value) => new TransformSummary.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (displayData != null) {
- _json["displayData"] = displayData.map((value) => (value).toJson()).toList();
+ _json["displayData"] =
+ displayData.map((value) => (value).toJson()).toList();
}
if (executionPipelineStage != null) {
- _json["executionPipelineStage"] = executionPipelineStage.map((value) => (value).toJson()).toList();
+ _json["executionPipelineStage"] =
+ executionPipelineStage.map((value) => (value).toJson()).toList();
}
if (originalPipelineTransform != null) {
- _json["originalPipelineTransform"] = originalPipelineTransform.map((value) => (value).toJson()).toList();
+ _json["originalPipelineTransform"] =
+ originalPipelineTransform.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * Position defines a position within a collection of data. The value
- * can be either the end position, a key (used with ordered
- * collections), a byte offset, or a record index.
- */
+/// Position defines a position within a collection of data. The value
+/// can be either the end position, a key (used with ordered
+/// collections), a byte offset, or a record index.
class Position {
- /** Position is a byte offset. */
+ /// Position is a byte offset.
core.String byteOffset;
- /** CloudPosition is a concat position. */
+
+ /// CloudPosition is a concat position.
ConcatPosition concatPosition;
- /**
- * Position is past all other positions. Also useful for the end
- * position of an unbounded range.
- */
+
+ /// Position is past all other positions. Also useful for the end
+ /// position of an unbounded range.
core.bool end;
- /** Position is a string key, ordered lexicographically. */
+
+ /// Position is a string key, ordered lexicographically.
core.String key;
- /** Position is a record index. */
+
+ /// Position is a record index.
core.String recordIndex;
- /**
- * CloudPosition is a base64 encoded BatchShufflePosition (with FIXED
- * sharding).
- */
+
+ /// CloudPosition is a base64 encoded BatchShufflePosition (with FIXED
+ /// sharding).
core.String shufflePosition;
Position();
@@ -5678,7 +5899,8 @@ class Position {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (byteOffset != null) {
_json["byteOffset"] = byteOffset;
}
@@ -5701,39 +5923,33 @@ class Position {
}
}
-/**
- * Identifies a pubsub location to use for transferring data into or
- * out of a streaming Dataflow job.
- */
+/// Identifies a pubsub location to use for transferring data into or
+/// out of a streaming Dataflow job.
class PubsubLocation {
- /** Indicates whether the pipeline allows late-arriving data. */
+ /// Indicates whether the pipeline allows late-arriving data.
core.bool dropLateData;
- /**
- * If set, contains a pubsub label from which to extract record ids.
- * If left empty, record deduplication will be strictly best effort.
- */
+
+ /// If set, contains a pubsub label from which to extract record ids.
+ /// If left empty, record deduplication will be strictly best effort.
core.String idLabel;
- /**
- * A pubsub subscription, in the form of
- * "pubsub.googleapis.com/subscriptions/<project-id>/<subscription-name>"
- */
+
+ /// A pubsub subscription, in the form of
+ /// "pubsub.googleapis.com/subscriptions/<project-id>/<subscription-name>"
core.String subscription;
- /**
- * If set, contains a pubsub label from which to extract record timestamps.
- * If left empty, record timestamps will be generated upon arrival.
- */
+
+ /// If set, contains a pubsub label from which to extract record timestamps.
+ /// If left empty, record timestamps will be generated upon arrival.
core.String timestampLabel;
- /**
- * A pubsub topic, in the form of
- * "pubsub.googleapis.com/topics/<project-id>/<topic-name>"
- */
+
+ /// A pubsub topic, in the form of
+ /// "pubsub.googleapis.com/topics/<project-id>/<topic-name>"
core.String topic;
- /**
- * If set, specifies the pubsub subscription that will be used for tracking
- * custom time timestamps for watermark estimation.
- */
+
+ /// If set, specifies the pubsub subscription that will be used for tracking
+ /// custom time timestamps for watermark estimation.
core.String trackingSubscription;
- /** If true, then the client has requested to get pubsub attributes. */
+
+ /// If true, then the client has requested to get pubsub attributes.
core.bool withAttributes;
PubsubLocation();
@@ -5763,7 +5979,8 @@ class PubsubLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dropLateData != null) {
_json["dropLateData"] = dropLateData;
}
@@ -5789,12 +6006,10 @@ class PubsubLocation {
}
}
-/**
- * An instruction that reads records.
- * Takes no inputs, produces one output.
- */
+/// An instruction that reads records.
+/// Takes no inputs, produces one output.
class ReadInstruction {
- /** The source to read from. */
+ /// The source to read from.
Source source;
ReadInstruction();
@@ -5806,7 +6021,8 @@ class ReadInstruction {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (source != null) {
_json["source"] = (source).toJson();
}
@@ -5814,24 +6030,23 @@ class ReadInstruction {
}
}
-/** Request to report the status of WorkItems. */
+/// Request to report the status of WorkItems.
class ReportWorkItemStatusRequest {
- /** The current timestamp at the worker. */
+ /// The current timestamp at the worker.
core.String currentWorkerTime;
- /** The location which contains the WorkItem's job. */
+
+ /// The location which contains the WorkItem's job.
core.String location;
- /**
- * The order is unimportant, except that the order of the
- * WorkItemServiceState messages in the ReportWorkItemStatusResponse
- * corresponds to the order of WorkItemStatus messages here.
- */
+
+ /// The order is unimportant, except that the order of the
+ /// WorkItemServiceState messages in the ReportWorkItemStatusResponse
+ /// corresponds to the order of WorkItemStatus messages here.
core.List<WorkItemStatus> workItemStatuses;
- /**
- * The ID of the worker reporting the WorkItem status. If this
- * does not match the ID of the worker which the Dataflow service
- * believes currently has the lease on the WorkItem, the report
- * will be dropped (with an error response).
- */
+
+ /// The ID of the worker reporting the WorkItem status. If this
+ /// does not match the ID of the worker which the Dataflow service
+ /// believes currently has the lease on the WorkItem, the report
+ /// will be dropped (with an error response).
core.String workerId;
ReportWorkItemStatusRequest();
@@ -5844,7 +6059,9 @@ class ReportWorkItemStatusRequest {
location = _json["location"];
}
if (_json.containsKey("workItemStatuses")) {
- workItemStatuses = _json["workItemStatuses"].map((value) => new WorkItemStatus.fromJson(value)).toList();
+ workItemStatuses = _json["workItemStatuses"]
+ .map((value) => new WorkItemStatus.fromJson(value))
+ .toList();
}
if (_json.containsKey("workerId")) {
workerId = _json["workerId"];
@@ -5852,7 +6069,8 @@ class ReportWorkItemStatusRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (currentWorkerTime != null) {
_json["currentWorkerTime"] = currentWorkerTime;
}
@@ -5860,7 +6078,8 @@ class ReportWorkItemStatusRequest {
_json["location"] = location;
}
if (workItemStatuses != null) {
- _json["workItemStatuses"] = workItemStatuses.map((value) => (value).toJson()).toList();
+ _json["workItemStatuses"] =
+ workItemStatuses.map((value) => (value).toJson()).toList();
}
if (workerId != null) {
_json["workerId"] = workerId;
@@ -5869,48 +6088,47 @@ class ReportWorkItemStatusRequest {
}
}
-/** Response from a request to report the status of WorkItems. */
+/// Response from a request to report the status of WorkItems.
class ReportWorkItemStatusResponse {
- /**
- * A set of messages indicating the service-side state for each
- * WorkItem whose status was reported, in the same order as the
- * WorkItemStatus messages in the ReportWorkItemStatusRequest which
- * resulting in this response.
- */
+ /// A set of messages indicating the service-side state for each
+ /// WorkItem whose status was reported, in the same order as the
+ /// WorkItemStatus messages in the ReportWorkItemStatusRequest which
+ /// resulting in this response.
core.List<WorkItemServiceState> workItemServiceStates;
ReportWorkItemStatusResponse();
ReportWorkItemStatusResponse.fromJson(core.Map _json) {
if (_json.containsKey("workItemServiceStates")) {
- workItemServiceStates = _json["workItemServiceStates"].map((value) => new WorkItemServiceState.fromJson(value)).toList();
+ workItemServiceStates = _json["workItemServiceStates"]
+ .map((value) => new WorkItemServiceState.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (workItemServiceStates != null) {
- _json["workItemServiceStates"] = workItemServiceStates.map((value) => (value).toJson()).toList();
+ _json["workItemServiceStates"] =
+ workItemServiceStates.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * Represents the level of parallelism in a WorkItem's input,
- * reported by the worker.
- */
+/// Represents the level of parallelism in a WorkItem's input,
+/// reported by the worker.
class ReportedParallelism {
- /**
- * Specifies whether the parallelism is infinite. If true, "value" is
- * ignored.
- * Infinite parallelism means the service will assume that the work item
- * can always be split into more non-empty work items by dynamic splitting.
- * This is a work-around for lack of support for infinity by the current
- * JSON-based Java RPC stack.
- */
+ /// Specifies whether the parallelism is infinite. If true, "value" is
+ /// ignored.
+ /// Infinite parallelism means the service will assume that the work item
+ /// can always be split into more non-empty work items by dynamic splitting.
+ /// This is a work-around for lack of support for infinity by the current
+ /// JSON-based Java RPC stack.
core.bool isInfinite;
- /** Specifies the level of parallelism in case it is finite. */
+
+ /// Specifies the level of parallelism in case it is finite.
core.double value;
ReportedParallelism();
@@ -5925,7 +6143,8 @@ class ReportedParallelism {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (isInfinite != null) {
_json["isInfinite"] = isInfinite;
}
@@ -5936,25 +6155,25 @@ class ReportedParallelism {
}
}
-/**
- * Worker metrics exported from workers. This contains resource utilization
- * metrics accumulated from a variety of sources. For more information, see
- * go/df-resource-signals.
- */
+/// Worker metrics exported from workers. This contains resource utilization
+/// metrics accumulated from a variety of sources. For more information, see
+/// go/df-resource-signals.
class ResourceUtilizationReport {
- /** CPU utilization samples. */
+ /// CPU utilization samples.
core.List<CPUTime> cpuTime;
ResourceUtilizationReport();
ResourceUtilizationReport.fromJson(core.Map _json) {
if (_json.containsKey("cpuTime")) {
- cpuTime = _json["cpuTime"].map((value) => new CPUTime.fromJson(value)).toList();
+ cpuTime =
+ _json["cpuTime"].map((value) => new CPUTime.fromJson(value)).toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (cpuTime != null) {
_json["cpuTime"] = cpuTime.map((value) => (value).toJson()).toList();
}
@@ -5962,49 +6181,43 @@ class ResourceUtilizationReport {
}
}
-/** Service-side response to WorkerMessage reporting resource utilization. */
+/// Service-side response to WorkerMessage reporting resource utilization.
class ResourceUtilizationReportResponse {
-
ResourceUtilizationReportResponse();
- ResourceUtilizationReportResponse.fromJson(core.Map _json) {
- }
+ ResourceUtilizationReportResponse.fromJson(core.Map _json) {}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
return _json;
}
}
-/** The environment values to set at runtime. */
+/// The environment values to set at runtime.
class RuntimeEnvironment {
- /**
- * Whether to bypass the safety checks for the job's temporary directory.
- * Use with caution.
- */
+ /// Whether to bypass the safety checks for the job's temporary directory.
+ /// Use with caution.
core.bool bypassTempDirValidation;
- /**
- * The machine type to use for the job. Defaults to the value from the
- * template if not specified.
- */
+
+ /// The machine type to use for the job. Defaults to the value from the
+ /// template if not specified.
core.String machineType;
- /**
- * The maximum number of Google Compute Engine instances to be made
- * available to your pipeline during execution, from 1 to 1000.
- */
+
+ /// The maximum number of Google Compute Engine instances to be made
+ /// available to your pipeline during execution, from 1 to 1000.
core.int maxWorkers;
- /** The email address of the service account to run the job as. */
+
+ /// The email address of the service account to run the job as.
core.String serviceAccountEmail;
- /**
- * The Cloud Storage path to use for temporary files.
- * Must be a valid Cloud Storage URL, beginning with `gs://`.
- */
+
+ /// The Cloud Storage path to use for temporary files.
+ /// Must be a valid Cloud Storage URL, beginning with `gs://`.
core.String tempLocation;
- /**
- * The Compute Engine [availability
- * zone](https://cloud.google.com/compute/docs/regions-zones/regions-zones)
- * for launching worker instances to run your pipeline.
- */
+
+ /// The Compute Engine [availability
+ /// zone](https://cloud.google.com/compute/docs/regions-zones/regions-zones)
+ /// for launching worker instances to run your pipeline.
core.String zone;
RuntimeEnvironment();
@@ -6031,7 +6244,8 @@ class RuntimeEnvironment {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (bypassTempDirValidation != null) {
_json["bypassTempDirValidation"] = bypassTempDirValidation;
}
@@ -6054,15 +6268,18 @@ class RuntimeEnvironment {
}
}
-/** Request to send encoded debug information. */
+/// Request to send encoded debug information.
class SendDebugCaptureRequest {
- /** The internal component id for which debug information is sent. */
+ /// The internal component id for which debug information is sent.
core.String componentId;
- /** The encoded debug information. */
+
+ /// The encoded debug information.
core.String data;
- /** The location which contains the job specified by job_id. */
+
+ /// The location which contains the job specified by job_id.
core.String location;
- /** The worker id, i.e., VM hostname. */
+
+ /// The worker id, i.e., VM hostname.
core.String workerId;
SendDebugCaptureRequest();
@@ -6083,7 +6300,8 @@ class SendDebugCaptureRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (componentId != null) {
_json["componentId"] = componentId;
}
@@ -6100,28 +6318,26 @@ class SendDebugCaptureRequest {
}
}
-/**
- * Response to a send capture request.
- * nothing
- */
+/// Response to a send capture request.
+/// nothing
class SendDebugCaptureResponse {
-
SendDebugCaptureResponse();
- SendDebugCaptureResponse.fromJson(core.Map _json) {
- }
+ SendDebugCaptureResponse.fromJson(core.Map _json) {}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
return _json;
}
}
-/** A request for sending worker messages to the service. */
+/// A request for sending worker messages to the service.
class SendWorkerMessagesRequest {
- /** The location which contains the job */
+ /// The location which contains the job
core.String location;
- /** The WorkerMessages to send. */
+
+ /// The WorkerMessages to send.
core.List<WorkerMessage> workerMessages;
SendWorkerMessagesRequest();
@@ -6131,81 +6347,92 @@ class SendWorkerMessagesRequest {
location = _json["location"];
}
if (_json.containsKey("workerMessages")) {
- workerMessages = _json["workerMessages"].map((value) => new WorkerMessage.fromJson(value)).toList();
+ workerMessages = _json["workerMessages"]
+ .map((value) => new WorkerMessage.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (location != null) {
_json["location"] = location;
}
if (workerMessages != null) {
- _json["workerMessages"] = workerMessages.map((value) => (value).toJson()).toList();
+ _json["workerMessages"] =
+ workerMessages.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** The response to the worker messages. */
+/// The response to the worker messages.
class SendWorkerMessagesResponse {
- /** The servers response to the worker messages. */
+ /// The servers response to the worker messages.
core.List<WorkerMessageResponse> workerMessageResponses;
SendWorkerMessagesResponse();
SendWorkerMessagesResponse.fromJson(core.Map _json) {
if (_json.containsKey("workerMessageResponses")) {
- workerMessageResponses = _json["workerMessageResponses"].map((value) => new WorkerMessageResponse.fromJson(value)).toList();
+ workerMessageResponses = _json["workerMessageResponses"]
+ .map((value) => new WorkerMessageResponse.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (workerMessageResponses != null) {
- _json["workerMessageResponses"] = workerMessageResponses.map((value) => (value).toJson()).toList();
+ _json["workerMessageResponses"] =
+ workerMessageResponses.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** Describes a particular function to invoke. */
+/// Describes a particular function to invoke.
class SeqMapTask {
- /** Information about each of the inputs. */
+ /// Information about each of the inputs.
core.List<SideInputInfo> inputs;
- /** The user-provided name of the SeqDo operation. */
+
+ /// The user-provided name of the SeqDo operation.
core.String name;
- /** Information about each of the outputs. */
+
+ /// Information about each of the outputs.
core.List<SeqMapTaskOutputInfo> outputInfos;
- /**
- * System-defined name of the stage containing the SeqDo operation.
- * Unique across the workflow.
- */
+
+ /// System-defined name of the stage containing the SeqDo operation.
+ /// Unique across the workflow.
core.String stageName;
- /**
- * System-defined name of the SeqDo operation.
- * Unique across the workflow.
- */
+
+ /// System-defined name of the SeqDo operation.
+ /// Unique across the workflow.
core.String systemName;
- /**
- * The user function to invoke.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The user function to invoke.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> userFn;
SeqMapTask();
SeqMapTask.fromJson(core.Map _json) {
if (_json.containsKey("inputs")) {
- inputs = _json["inputs"].map((value) => new SideInputInfo.fromJson(value)).toList();
+ inputs = _json["inputs"]
+ .map((value) => new SideInputInfo.fromJson(value))
+ .toList();
}
if (_json.containsKey("name")) {
name = _json["name"];
}
if (_json.containsKey("outputInfos")) {
- outputInfos = _json["outputInfos"].map((value) => new SeqMapTaskOutputInfo.fromJson(value)).toList();
+ outputInfos = _json["outputInfos"]
+ .map((value) => new SeqMapTaskOutputInfo.fromJson(value))
+ .toList();
}
if (_json.containsKey("stageName")) {
stageName = _json["stageName"];
@@ -6219,7 +6446,8 @@ class SeqMapTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (inputs != null) {
_json["inputs"] = inputs.map((value) => (value).toJson()).toList();
}
@@ -6227,7 +6455,8 @@ class SeqMapTask {
_json["name"] = name;
}
if (outputInfos != null) {
- _json["outputInfos"] = outputInfos.map((value) => (value).toJson()).toList();
+ _json["outputInfos"] =
+ outputInfos.map((value) => (value).toJson()).toList();
}
if (stageName != null) {
_json["stageName"] = stageName;
@@ -6242,11 +6471,12 @@ class SeqMapTask {
}
}
-/** Information about an output of a SeqMapTask. */
+/// Information about an output of a SeqMapTask.
class SeqMapTaskOutputInfo {
- /** The sink to write the output value to. */
+ /// The sink to write the output value to.
Sink sink;
- /** The id of the TupleTag the user code will tag the output value by. */
+
+ /// The id of the TupleTag the user code will tag the output value by.
core.String tag;
SeqMapTaskOutputInfo();
@@ -6261,7 +6491,8 @@ class SeqMapTaskOutputInfo {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (sink != null) {
_json["sink"] = (sink).toJson();
}
@@ -6272,11 +6503,12 @@ class SeqMapTaskOutputInfo {
}
}
-/** A task which consists of a shell command for the worker to execute. */
+/// A task which consists of a shell command for the worker to execute.
class ShellTask {
- /** The shell command to run. */
+ /// The shell command to run.
core.String command;
- /** Exit code for the task. */
+
+ /// Exit code for the task.
core.int exitCode;
ShellTask();
@@ -6291,7 +6523,8 @@ class ShellTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (command != null) {
_json["command"] = command;
}
@@ -6302,26 +6535,55 @@ class ShellTask {
}
}
-/** Information about a side input of a DoFn or an input of a SeqDoFn. */
+/// Uniquely identifies a side input.
+class SideInputId {
+ /// The step that receives and usually consumes this side input.
+ core.String declaringStepName;
+
+ /// The index of the side input, from the list of non_parallel_inputs.
+ core.int inputIndex;
+
+ SideInputId();
+
+ SideInputId.fromJson(core.Map _json) {
+ if (_json.containsKey("declaringStepName")) {
+ declaringStepName = _json["declaringStepName"];
+ }
+ if (_json.containsKey("inputIndex")) {
+ inputIndex = _json["inputIndex"];
+ }
+ }
+
+ core.Map<core.String, core.Object> toJson() {
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
+ if (declaringStepName != null) {
+ _json["declaringStepName"] = declaringStepName;
+ }
+ if (inputIndex != null) {
+ _json["inputIndex"] = inputIndex;
+ }
+ return _json;
+ }
+}
+
+/// Information about a side input of a DoFn or an input of a SeqDoFn.
class SideInputInfo {
- /**
- * How to interpret the source element(s) as a side input value.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// How to interpret the source element(s) as a side input value.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> kind;
- /**
- * The source(s) to read element(s) from to get the value of this side input.
- * If more than one source, then the elements are taken from the
- * sources, in the specified order if order matters.
- * At least one source is required.
- */
+
+ /// The source(s) to read element(s) from to get the value of this side
+ /// input.
+ /// If more than one source, then the elements are taken from the
+ /// sources, in the specified order if order matters.
+ /// At least one source is required.
core.List<Source> sources;
- /**
- * The id of the tag the user code will access this side input by;
- * this should correspond to the tag of some MultiOutputInfo.
- */
+
+ /// The id of the tag the user code will access this side input by;
+ /// this should correspond to the tag of some MultiOutputInfo.
core.String tag;
SideInputInfo();
@@ -6331,7 +6593,8 @@ class SideInputInfo {
kind = _json["kind"];
}
if (_json.containsKey("sources")) {
- sources = _json["sources"].map((value) => new Source.fromJson(value)).toList();
+ sources =
+ _json["sources"].map((value) => new Source.fromJson(value)).toList();
}
if (_json.containsKey("tag")) {
tag = _json["tag"];
@@ -6339,7 +6602,8 @@ class SideInputInfo {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (kind != null) {
_json["kind"] = kind;
}
@@ -6353,21 +6617,18 @@ class SideInputInfo {
}
}
-/** A sink that records can be encoded and written to. */
+/// A sink that records can be encoded and written to.
class Sink {
- /**
- * The codec to use to encode data written to the sink.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// The codec to use to encode data written to the sink.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> codec;
- /**
- * The sink to write to, plus its parameters.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The sink to write to, plus its parameters.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> spec;
Sink();
@@ -6382,7 +6643,8 @@ class Sink {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (codec != null) {
_json["codec"] = codec;
}
@@ -6393,65 +6655,59 @@ class Sink {
}
}
-/** A source that records can be read and decoded from. */
+/// A source that records can be read and decoded from.
class Source {
- /**
- * While splitting, sources may specify the produced bundles
- * as differences against another source, in order to save backend-side
- * memory and allow bigger jobs. For details, see SourceSplitRequest.
- * To support this use case, the full set of parameters of the source
- * is logically obtained by taking the latest explicitly specified value
- * of each parameter in the order:
- * base_specs (later items win), spec (overrides anything in base_specs).
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// While splitting, sources may specify the produced bundles
+ /// as differences against another source, in order to save backend-side
+ /// memory and allow bigger jobs. For details, see SourceSplitRequest.
+ /// To support this use case, the full set of parameters of the source
+ /// is logically obtained by taking the latest explicitly specified value
+ /// of each parameter in the order:
+ /// base_specs (later items win), spec (overrides anything in base_specs).
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.List<core.Map<core.String, core.Object>> baseSpecs;
- /**
- * The codec to use to decode data read from the source.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The codec to use to decode data read from the source.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> codec;
- /**
- * Setting this value to true hints to the framework that the source
- * doesn't need splitting, and using SourceSplitRequest on it would
- * yield SOURCE_SPLIT_OUTCOME_USE_CURRENT.
- *
- * E.g. a file splitter may set this to true when splitting a single file
- * into a set of byte ranges of appropriate size, and set this
- * to false when splitting a filepattern into individual files.
- * However, for efficiency, a file splitter may decide to produce
- * file subranges directly from the filepattern to avoid a splitting
- * round-trip.
- *
- * See SourceSplitRequest for an overview of the splitting process.
- *
- * This field is meaningful only in the Source objects populated
- * by the user (e.g. when filling in a DerivedSource).
- * Source objects supplied by the framework to the user don't have
- * this field populated.
- */
+
+ /// Setting this value to true hints to the framework that the source
+ /// doesn't need splitting, and using SourceSplitRequest on it would
+ /// yield SOURCE_SPLIT_OUTCOME_USE_CURRENT.
+ ///
+ /// E.g. a file splitter may set this to true when splitting a single file
+ /// into a set of byte ranges of appropriate size, and set this
+ /// to false when splitting a filepattern into individual files.
+ /// However, for efficiency, a file splitter may decide to produce
+ /// file subranges directly from the filepattern to avoid a splitting
+ /// round-trip.
+ ///
+ /// See SourceSplitRequest for an overview of the splitting process.
+ ///
+ /// This field is meaningful only in the Source objects populated
+ /// by the user (e.g. when filling in a DerivedSource).
+ /// Source objects supplied by the framework to the user don't have
+ /// this field populated.
core.bool doesNotNeedSplitting;
- /**
- * Optionally, metadata for this source can be supplied right away,
- * avoiding a SourceGetMetadataOperation roundtrip
- * (see SourceOperationRequest).
- *
- * This field is meaningful only in the Source objects populated
- * by the user (e.g. when filling in a DerivedSource).
- * Source objects supplied by the framework to the user don't have
- * this field populated.
- */
+
+ /// Optionally, metadata for this source can be supplied right away,
+ /// avoiding a SourceGetMetadataOperation roundtrip
+ /// (see SourceOperationRequest).
+ ///
+ /// This field is meaningful only in the Source objects populated
+ /// by the user (e.g. when filling in a DerivedSource).
+ /// Source objects supplied by the framework to the user don't have
+ /// this field populated.
SourceMetadata metadata;
- /**
- * The source to read from, plus its parameters.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The source to read from, plus its parameters.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> spec;
Source();
@@ -6475,7 +6731,8 @@ class Source {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (baseSpecs != null) {
_json["baseSpecs"] = baseSpecs;
}
@@ -6495,15 +6752,18 @@ class Source {
}
}
-/** DEPRECATED in favor of DynamicSourceSplit. */
+/// DEPRECATED in favor of DynamicSourceSplit.
class SourceFork {
- /** DEPRECATED */
+ /// DEPRECATED
SourceSplitShard primary;
- /** DEPRECATED */
+
+ /// DEPRECATED
DerivedSource primarySource;
- /** DEPRECATED */
+
+ /// DEPRECATED
SourceSplitShard residual;
- /** DEPRECATED */
+
+ /// DEPRECATED
DerivedSource residualSource;
SourceFork();
@@ -6524,7 +6784,8 @@ class SourceFork {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (primary != null) {
_json["primary"] = (primary).toJson();
}
@@ -6541,9 +6802,9 @@ class SourceFork {
}
}
-/** A request to compute the SourceMetadata of a Source. */
+/// A request to compute the SourceMetadata of a Source.
class SourceGetMetadataRequest {
- /** Specification of the source whose metadata should be computed. */
+ /// Specification of the source whose metadata should be computed.
Source source;
SourceGetMetadataRequest();
@@ -6555,7 +6816,8 @@ class SourceGetMetadataRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (source != null) {
_json["source"] = (source).toJson();
}
@@ -6563,9 +6825,9 @@ class SourceGetMetadataRequest {
}
}
-/** The result of a SourceGetMetadataOperation. */
+/// The result of a SourceGetMetadataOperation.
class SourceGetMetadataResponse {
- /** The computed metadata. */
+ /// The computed metadata.
SourceMetadata metadata;
SourceGetMetadataResponse();
@@ -6577,7 +6839,8 @@ class SourceGetMetadataResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (metadata != null) {
_json["metadata"] = (metadata).toJson();
}
@@ -6585,26 +6848,20 @@ class SourceGetMetadataResponse {
}
}
-/**
- * Metadata about a Source useful for automatically optimizing
- * and tuning the pipeline, etc.
- */
+/// Metadata about a Source useful for automatically optimizing
+/// and tuning the pipeline, etc.
class SourceMetadata {
- /**
- * An estimate of the total size (in bytes) of the data that would be
- * read from this source. This estimate is in terms of external storage
- * size, before any decompression or other processing done by the reader.
- */
+ /// An estimate of the total size (in bytes) of the data that would be
+ /// read from this source. This estimate is in terms of external storage
+ /// size, before any decompression or other processing done by the reader.
core.String estimatedSizeBytes;
- /**
- * Specifies that the size of this source is known to be infinite
- * (this is a streaming source).
- */
+
+ /// Specifies that the size of this source is known to be infinite
+ /// (this is a streaming source).
core.bool infinite;
- /**
- * Whether this source is known to produce key/value pairs with
- * the (encoded) keys in lexicographically sorted order.
- */
+
+ /// Whether this source is known to produce key/value pairs with
+ /// the (encoded) keys in lexicographically sorted order.
core.bool producesSortedKeys;
SourceMetadata();
@@ -6622,7 +6879,8 @@ class SourceMetadata {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (estimatedSizeBytes != null) {
_json["estimatedSizeBytes"] = estimatedSizeBytes;
}
@@ -6636,14 +6894,13 @@ class SourceMetadata {
}
}
-/**
- * A work item that represents the different operations that can be
- * performed on a user-defined Source specification.
- */
+/// A work item that represents the different operations that can be
+/// performed on a user-defined Source specification.
class SourceOperationRequest {
- /** Information about a request to get metadata about a source. */
+ /// Information about a request to get metadata about a source.
SourceGetMetadataRequest getMetadata;
- /** Information about a request to split a source. */
+
+ /// Information about a request to split a source.
SourceSplitRequest split;
SourceOperationRequest();
@@ -6658,7 +6915,8 @@ class SourceOperationRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (getMetadata != null) {
_json["getMetadata"] = (getMetadata).toJson();
}
@@ -6669,22 +6927,22 @@ class SourceOperationRequest {
}
}
-/**
- * The result of a SourceOperationRequest, specified in
- * ReportWorkItemStatusRequest.source_operation when the work item
- * is completed.
- */
+/// The result of a SourceOperationRequest, specified in
+/// ReportWorkItemStatusRequest.source_operation when the work item
+/// is completed.
class SourceOperationResponse {
- /** A response to a request to get metadata about a source. */
+ /// A response to a request to get metadata about a source.
SourceGetMetadataResponse getMetadata;
- /** A response to a request to split a source. */
+
+ /// A response to a request to split a source.
SourceSplitResponse split;
SourceOperationResponse();
SourceOperationResponse.fromJson(core.Map _json) {
if (_json.containsKey("getMetadata")) {
- getMetadata = new SourceGetMetadataResponse.fromJson(_json["getMetadata"]);
+ getMetadata =
+ new SourceGetMetadataResponse.fromJson(_json["getMetadata"]);
}
if (_json.containsKey("split")) {
split = new SourceSplitResponse.fromJson(_json["split"]);
@@ -6692,7 +6950,8 @@ class SourceOperationResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (getMetadata != null) {
_json["getMetadata"] = (getMetadata).toJson();
}
@@ -6703,17 +6962,14 @@ class SourceOperationResponse {
}
}
-/**
- * Hints for splitting a Source into bundles (parts for parallel
- * processing) using SourceSplitRequest.
- */
+/// Hints for splitting a Source into bundles (parts for parallel
+/// processing) using SourceSplitRequest.
class SourceSplitOptions {
- /**
- * The source should be split into a set of bundles where the estimated size
- * of each is approximately this many bytes.
- */
+ /// The source should be split into a set of bundles where the estimated size
+ /// of each is approximately this many bytes.
core.String desiredBundleSizeBytes;
- /** DEPRECATED in favor of desired_bundle_size_bytes. */
+
+ /// DEPRECATED in favor of desired_bundle_size_bytes.
core.String desiredShardSizeBytes;
SourceSplitOptions();
@@ -6728,7 +6984,8 @@ class SourceSplitOptions {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (desiredBundleSizeBytes != null) {
_json["desiredBundleSizeBytes"] = desiredBundleSizeBytes;
}
@@ -6739,26 +6996,27 @@ class SourceSplitOptions {
}
}
-/**
- * Represents the operation to split a high-level Source specification
- * into bundles (parts for parallel processing).
- *
- * At a high level, splitting of a source into bundles happens as follows:
- * SourceSplitRequest is applied to the source. If it returns
- * SOURCE_SPLIT_OUTCOME_USE_CURRENT, no further splitting happens and the source
- * is used "as is". Otherwise, splitting is applied recursively to each
- * produced DerivedSource.
- *
- * As an optimization, for any Source, if its does_not_need_splitting is
- * true, the framework assumes that splitting this source would return
- * SOURCE_SPLIT_OUTCOME_USE_CURRENT, and doesn't initiate a SourceSplitRequest.
- * This applies both to the initial source being split and to bundles
- * produced from it.
- */
+/// Represents the operation to split a high-level Source specification
+/// into bundles (parts for parallel processing).
+///
+/// At a high level, splitting of a source into bundles happens as follows:
+/// SourceSplitRequest is applied to the source. If it returns
+/// SOURCE_SPLIT_OUTCOME_USE_CURRENT, no further splitting happens and the
+/// source
+/// is used "as is". Otherwise, splitting is applied recursively to each
+/// produced DerivedSource.
+///
+/// As an optimization, for any Source, if its does_not_need_splitting is
+/// true, the framework assumes that splitting this source would return
+/// SOURCE_SPLIT_OUTCOME_USE_CURRENT, and doesn't initiate a
+/// SourceSplitRequest.
+/// This applies both to the initial source being split and to bundles
+/// produced from it.
class SourceSplitRequest {
- /** Hints for tuning the splitting process. */
+ /// Hints for tuning the splitting process.
SourceSplitOptions options;
- /** Specification of the source to be split. */
+
+ /// Specification of the source to be split.
Source source;
SourceSplitRequest();
@@ -6773,7 +7031,8 @@ class SourceSplitRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (options != null) {
_json["options"] = (options).toJson();
}
@@ -6784,48 +7043,52 @@ class SourceSplitRequest {
}
}
-/** The response to a SourceSplitRequest. */
+/// The response to a SourceSplitRequest.
class SourceSplitResponse {
- /**
- * If outcome is SPLITTING_HAPPENED, then this is a list of bundles
- * into which the source was split. Otherwise this field is ignored.
- * This list can be empty, which means the source represents an empty input.
- */
+ /// If outcome is SPLITTING_HAPPENED, then this is a list of bundles
+ /// into which the source was split. Otherwise this field is ignored.
+ /// This list can be empty, which means the source represents an empty input.
core.List<DerivedSource> bundles;
- /**
- * Indicates whether splitting happened and produced a list of bundles.
- * If this is USE_CURRENT_SOURCE_AS_IS, the current source should
- * be processed "as is" without splitting. "bundles" is ignored in this case.
- * If this is SPLITTING_HAPPENED, then "bundles" contains a list of
- * bundles into which the source was split.
- * Possible string values are:
- * - "SOURCE_SPLIT_OUTCOME_UNKNOWN" : The source split outcome is unknown, or
- * unspecified.
- * - "SOURCE_SPLIT_OUTCOME_USE_CURRENT" : The current source should be
- * processed "as is" without splitting.
- * - "SOURCE_SPLIT_OUTCOME_SPLITTING_HAPPENED" : Splitting produced a list of
- * bundles.
- */
+
+ /// Indicates whether splitting happened and produced a list of bundles.
+ /// If this is USE_CURRENT_SOURCE_AS_IS, the current source should
+ /// be processed "as is" without splitting. "bundles" is ignored in this
+ /// case.
+ /// If this is SPLITTING_HAPPENED, then "bundles" contains a list of
+ /// bundles into which the source was split.
+ /// Possible string values are:
+ /// - "SOURCE_SPLIT_OUTCOME_UNKNOWN" : The source split outcome is unknown,
+ /// or unspecified.
+ /// - "SOURCE_SPLIT_OUTCOME_USE_CURRENT" : The current source should be
+ /// processed "as is" without splitting.
+ /// - "SOURCE_SPLIT_OUTCOME_SPLITTING_HAPPENED" : Splitting produced a list
+ /// of bundles.
core.String outcome;
- /** DEPRECATED in favor of bundles. */
+
+ /// DEPRECATED in favor of bundles.
core.List<SourceSplitShard> shards;
SourceSplitResponse();
SourceSplitResponse.fromJson(core.Map _json) {
if (_json.containsKey("bundles")) {
- bundles = _json["bundles"].map((value) => new DerivedSource.fromJson(value)).toList();
+ bundles = _json["bundles"]
+ .map((value) => new DerivedSource.fromJson(value))
+ .toList();
}
if (_json.containsKey("outcome")) {
outcome = _json["outcome"];
}
if (_json.containsKey("shards")) {
- shards = _json["shards"].map((value) => new SourceSplitShard.fromJson(value)).toList();
+ shards = _json["shards"]
+ .map((value) => new SourceSplitShard.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (bundles != null) {
_json["bundles"] = bundles.map((value) => (value).toJson()).toList();
}
@@ -6839,22 +7102,21 @@ class SourceSplitResponse {
}
}
-/** DEPRECATED in favor of DerivedSource. */
+/// DEPRECATED in favor of DerivedSource.
class SourceSplitShard {
- /**
- * DEPRECATED
- * Possible string values are:
- * - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or
- * unspecified.
- * - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent
- * Source with no base.
- * - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on the
- * Source being split.
- * - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on
- * the base of the Source being split.
- */
+ /// DEPRECATED
+ /// Possible string values are:
+ /// - "SOURCE_DERIVATION_MODE_UNKNOWN" : The source derivation is unknown, or
+ /// unspecified.
+ /// - "SOURCE_DERIVATION_MODE_INDEPENDENT" : Produce a completely independent
+ /// Source with no base.
+ /// - "SOURCE_DERIVATION_MODE_CHILD_OF_CURRENT" : Produce a Source based on
+ /// the Source being split.
+ /// - "SOURCE_DERIVATION_MODE_SIBLING_OF_CURRENT" : Produce a Source based on
+ /// the base of the Source being split.
core.String derivationMode;
- /** DEPRECATED */
+
+ /// DEPRECATED
Source source;
SourceSplitShard();
@@ -6869,7 +7131,8 @@ class SourceSplitShard {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (derivationMode != null) {
_json["derivationMode"] = derivationMode;
}
@@ -6880,14 +7143,13 @@ class SourceSplitShard {
}
}
-/**
- * A representation of an int64, n, that is immune to precision loss when
- * encoded in JSON.
- */
+/// A representation of an int64, n, that is immune to precision loss when
+/// encoded in JSON.
class SplitInt64 {
- /** The high order bits, including the sign: n >> 32. */
+ /// The high order bits, including the sign: n >> 32.
core.int highBits;
- /** The low order bits: n & 0xffffffff. */
+
+ /// The low order bits: n & 0xffffffff.
core.int lowBits;
SplitInt64();
@@ -6902,7 +7164,8 @@ class SplitInt64 {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (highBits != null) {
_json["highBits"] = highBits;
}
@@ -6913,18 +7176,19 @@ class SplitInt64 {
}
}
-/** Description of an input or output of an execution stage. */
+/// Description of an input or output of an execution stage.
class StageSource {
- /** Dataflow service generated name for this source. */
+ /// Dataflow service generated name for this source.
core.String name;
- /**
- * User name for the original user transform or collection with which this
- * source is most closely associated.
- */
+
+ /// User name for the original user transform or collection with which this
+ /// source is most closely associated.
core.String originalTransformOrCollection;
- /** Size of the source, if measurable. */
+
+ /// Size of the source, if measurable.
core.String sizeBytes;
- /** Human-readable name for this source; may be user or system generated. */
+
+ /// Human-readable name for this source; may be user or system generated.
core.String userName;
StageSource();
@@ -6945,7 +7209,8 @@ class StageSource {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (name != null) {
_json["name"] = name;
}
@@ -6962,11 +7227,12 @@ class StageSource {
}
}
-/** State family configuration. */
+/// State family configuration.
class StateFamilyConfig {
- /** If true, this family corresponds to a read operation. */
+ /// If true, this family corresponds to a read operation.
core.bool isRead;
- /** The state family value. */
+
+ /// The state family value.
core.String stateFamily;
StateFamilyConfig();
@@ -6981,7 +7247,8 @@ class StateFamilyConfig {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (isRead != null) {
_json["isRead"] = isRead;
}
@@ -6992,78 +7259,81 @@ class StateFamilyConfig {
}
}
-/**
- * The `Status` type defines a logical error model that is suitable for
- * different
- * programming environments, including REST APIs and RPC APIs. It is used by
- * [gRPC](https://github.com/grpc). The error model is designed to be:
- *
- * - Simple to use and understand for most users
- * - Flexible enough to meet unexpected needs
- *
- * # Overview
- *
- * The `Status` message contains three pieces of data: error code, error
- * message,
- * and error details. The error code should be an enum value of
- * google.rpc.Code, but it may accept additional error codes if needed. The
- * error message should be a developer-facing English message that helps
- * developers *understand* and *resolve* the error. If a localized user-facing
- * error message is needed, put the localized message in the error details or
- * localize it in the client. The optional error details may contain arbitrary
- * information about the error. There is a predefined set of error detail types
- * in the package `google.rpc` that can be used for common error conditions.
- *
- * # Language mapping
- *
- * The `Status` message is the logical representation of the error model, but it
- * is not necessarily the actual wire format. When the `Status` message is
- * exposed in different client libraries and different wire protocols, it can be
- * mapped differently. For example, it will likely be mapped to some exceptions
- * in Java, but more likely mapped to some error codes in C.
- *
- * # Other uses
- *
- * The error model and the `Status` message can be used in a variety of
- * environments, either with or without APIs, to provide a
- * consistent developer experience across different environments.
- *
- * Example uses of this error model include:
- *
- * - Partial errors. If a service needs to return partial errors to the client,
- * it may embed the `Status` in the normal response to indicate the partial
- * errors.
- *
- * - Workflow errors. A typical workflow has multiple steps. Each step may
- * have a `Status` message for error reporting.
- *
- * - Batch operations. If a client uses batch request and batch response, the
- * `Status` message should be used directly inside batch response, one for
- * each error sub-response.
- *
- * - Asynchronous operations. If an API call embeds asynchronous operation
- * results in its response, the status of those operations should be
- * represented directly using the `Status` message.
- *
- * - Logging. If some API errors are stored in logs, the message `Status` could
- * be used directly after any stripping needed for security/privacy reasons.
- */
+/// The `Status` type defines a logical error model that is suitable for
+/// different
+/// programming environments, including REST APIs and RPC APIs. It is used by
+/// [gRPC](https://github.com/grpc). The error model is designed to be:
+///
+/// - Simple to use and understand for most users
+/// - Flexible enough to meet unexpected needs
+///
+/// # Overview
+///
+/// The `Status` message contains three pieces of data: error code, error
+/// message,
+/// and error details. The error code should be an enum value of
+/// google.rpc.Code, but it may accept additional error codes if needed. The
+/// error message should be a developer-facing English message that helps
+/// developers *understand* and *resolve* the error. If a localized user-facing
+/// error message is needed, put the localized message in the error details or
+/// localize it in the client. The optional error details may contain arbitrary
+/// information about the error. There is a predefined set of error detail
+/// types
+/// in the package `google.rpc` that can be used for common error conditions.
+///
+/// # Language mapping
+///
+/// The `Status` message is the logical representation of the error model, but
+/// it
+/// is not necessarily the actual wire format. When the `Status` message is
+/// exposed in different client libraries and different wire protocols, it can
+/// be
+/// mapped differently. For example, it will likely be mapped to some
+/// exceptions
+/// in Java, but more likely mapped to some error codes in C.
+///
+/// # Other uses
+///
+/// The error model and the `Status` message can be used in a variety of
+/// environments, either with or without APIs, to provide a
+/// consistent developer experience across different environments.
+///
+/// Example uses of this error model include:
+///
+/// - Partial errors. If a service needs to return partial errors to the
+/// client,
+/// it may embed the `Status` in the normal response to indicate the partial
+/// errors.
+///
+/// - Workflow errors. A typical workflow has multiple steps. Each step may
+/// have a `Status` message for error reporting.
+///
+/// - Batch operations. If a client uses batch request and batch response, the
+/// `Status` message should be used directly inside batch response, one for
+/// each error sub-response.
+///
+/// - Asynchronous operations. If an API call embeds asynchronous operation
+/// results in its response, the status of those operations should be
+/// represented directly using the `Status` message.
+///
+/// - Logging. If some API errors are stored in logs, the message `Status`
+/// could
+/// be used directly after any stripping needed for security/privacy reasons.
class Status {
- /** The status code, which should be an enum value of google.rpc.Code. */
+ /// The status code, which should be an enum value of google.rpc.Code.
core.int code;
- /**
- * A list of messages that carry the error details. There is a common set of
- * message types for APIs to use.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A list of messages that carry the error details. There is a common set
+ /// of
+ /// message types for APIs to use.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.List<core.Map<core.String, core.Object>> details;
- /**
- * A developer-facing error message, which should be in English. Any
- * user-facing error message should be localized and sent in the
- * google.rpc.Status.details field, or localized by the client.
- */
+
+ /// A developer-facing error message, which should be in English. Any
+ /// user-facing error message should be localized and sent in the
+ /// google.rpc.Status.details field, or localized by the client.
core.String message;
Status();
@@ -7081,7 +7351,8 @@ class Status {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (code != null) {
_json["code"] = code;
}
@@ -7095,49 +7366,45 @@ class Status {
}
}
-/**
- * Defines a particular step within a Cloud Dataflow job.
- *
- * A job consists of multiple steps, each of which performs some
- * specific operation as part of the overall job. Data is typically
- * passed from one step to another as part of the job.
- *
- * Here's an example of a sequence of steps which together implement a
- * Map-Reduce job:
- *
- * * Read a collection of data from some source, parsing the
- * collection's elements.
- *
- * * Validate the elements.
- *
- * * Apply a user-defined function to map each element to some value
- * and extract an element-specific key value.
- *
- * * Group elements with the same key into a single element with
- * that key, transforming a multiply-keyed collection into a
- * uniquely-keyed collection.
- *
- * * Write the elements out to some data sink.
- *
- * Note that the Cloud Dataflow service may be used to run many different
- * types of jobs, not just Map-Reduce.
- */
+/// Defines a particular step within a Cloud Dataflow job.
+///
+/// A job consists of multiple steps, each of which performs some
+/// specific operation as part of the overall job. Data is typically
+/// passed from one step to another as part of the job.
+///
+/// Here's an example of a sequence of steps which together implement a
+/// Map-Reduce job:
+///
+/// * Read a collection of data from some source, parsing the
+/// collection's elements.
+///
+/// * Validate the elements.
+///
+/// * Apply a user-defined function to map each element to some value
+/// and extract an element-specific key value.
+///
+/// * Group elements with the same key into a single element with
+/// that key, transforming a multiply-keyed collection into a
+/// uniquely-keyed collection.
+///
+/// * Write the elements out to some data sink.
+///
+/// Note that the Cloud Dataflow service may be used to run many different
+/// types of jobs, not just Map-Reduce.
class Step {
- /** The kind of step in the Cloud Dataflow job. */
+ /// The kind of step in the Cloud Dataflow job.
core.String kind;
- /**
- * The name that identifies the step. This must be unique for each
- * step with respect to all other steps in the Cloud Dataflow job.
- */
+
+ /// The name that identifies the step. This must be unique for each
+ /// step with respect to all other steps in the Cloud Dataflow job.
core.String name;
- /**
- * Named properties associated with the step. Each kind of
- * predefined step has its own required set of properties.
- * Must be provided on Create. Only retrieved with JOB_VIEW_ALL.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Named properties associated with the step. Each kind of
+ /// predefined step has its own required set of properties.
+ /// Must be provided on Create. Only retrieved with JOB_VIEW_ALL.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> properties;
Step();
@@ -7155,7 +7422,8 @@ class Step {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (kind != null) {
_json["kind"] = kind;
}
@@ -7169,42 +7437,45 @@ class Step {
}
}
-/**
- * Describes a stream of data, either as input to be processed or as
- * output of a streaming Dataflow job.
- */
+/// Describes a stream of data, either as input to be processed or as
+/// output of a streaming Dataflow job.
class StreamLocation {
- /** The stream is a custom source. */
+ /// The stream is a custom source.
CustomSourceLocation customSourceLocation;
- /** The stream is a pubsub stream. */
+
+ /// The stream is a pubsub stream.
PubsubLocation pubsubLocation;
- /** The stream is a streaming side input. */
+
+ /// The stream is a streaming side input.
StreamingSideInputLocation sideInputLocation;
- /**
- * The stream is part of another computation within the current
- * streaming Dataflow job.
- */
+
+ /// The stream is part of another computation within the current
+ /// streaming Dataflow job.
StreamingStageLocation streamingStageLocation;
StreamLocation();
StreamLocation.fromJson(core.Map _json) {
if (_json.containsKey("customSourceLocation")) {
- customSourceLocation = new CustomSourceLocation.fromJson(_json["customSourceLocation"]);
+ customSourceLocation =
+ new CustomSourceLocation.fromJson(_json["customSourceLocation"]);
}
if (_json.containsKey("pubsubLocation")) {
pubsubLocation = new PubsubLocation.fromJson(_json["pubsubLocation"]);
}
if (_json.containsKey("sideInputLocation")) {
- sideInputLocation = new StreamingSideInputLocation.fromJson(_json["sideInputLocation"]);
+ sideInputLocation =
+ new StreamingSideInputLocation.fromJson(_json["sideInputLocation"]);
}
if (_json.containsKey("streamingStageLocation")) {
- streamingStageLocation = new StreamingStageLocation.fromJson(_json["streamingStageLocation"]);
+ streamingStageLocation =
+ new StreamingStageLocation.fromJson(_json["streamingStageLocation"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (customSourceLocation != null) {
_json["customSourceLocation"] = (customSourceLocation).toJson();
}
@@ -7221,15 +7492,18 @@ class StreamLocation {
}
}
-/** Configuration information for a single streaming computation. */
+/// Configuration information for a single streaming computation.
class StreamingComputationConfig {
- /** Unique identifier for this computation. */
+ /// Unique identifier for this computation.
core.String computationId;
- /** Instructions that comprise the computation. */
+
+ /// Instructions that comprise the computation.
core.List<ParallelInstruction> instructions;
- /** Stage name of this computation. */
+
+ /// Stage name of this computation.
core.String stageName;
- /** System defined name for this computation. */
+
+ /// System defined name for this computation.
core.String systemName;
StreamingComputationConfig();
@@ -7239,7 +7513,9 @@ class StreamingComputationConfig {
computationId = _json["computationId"];
}
if (_json.containsKey("instructions")) {
- instructions = _json["instructions"].map((value) => new ParallelInstruction.fromJson(value)).toList();
+ instructions = _json["instructions"]
+ .map((value) => new ParallelInstruction.fromJson(value))
+ .toList();
}
if (_json.containsKey("stageName")) {
stageName = _json["stageName"];
@@ -7250,12 +7526,14 @@ class StreamingComputationConfig {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (computationId != null) {
_json["computationId"] = computationId;
}
if (instructions != null) {
- _json["instructions"] = instructions.map((value) => (value).toJson()).toList();
+ _json["instructions"] =
+ instructions.map((value) => (value).toJson()).toList();
}
if (stageName != null) {
_json["stageName"] = stageName;
@@ -7267,14 +7545,14 @@ class StreamingComputationConfig {
}
}
-/**
- * Describes full or partial data disk assignment information of the computation
- * ranges.
- */
+/// Describes full or partial data disk assignment information of the
+/// computation
+/// ranges.
class StreamingComputationRanges {
- /** The ID of the computation. */
+ /// The ID of the computation.
core.String computationId;
- /** Data disk assignments for ranges from this computation. */
+
+ /// Data disk assignments for ranges from this computation.
core.List<KeyRangeDataDiskAssignment> rangeAssignments;
StreamingComputationRanges();
@@ -7284,51 +7562,57 @@ class StreamingComputationRanges {
computationId = _json["computationId"];
}
if (_json.containsKey("rangeAssignments")) {
- rangeAssignments = _json["rangeAssignments"].map((value) => new KeyRangeDataDiskAssignment.fromJson(value)).toList();
+ rangeAssignments = _json["rangeAssignments"]
+ .map((value) => new KeyRangeDataDiskAssignment.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (computationId != null) {
_json["computationId"] = computationId;
}
if (rangeAssignments != null) {
- _json["rangeAssignments"] = rangeAssignments.map((value) => (value).toJson()).toList();
+ _json["rangeAssignments"] =
+ rangeAssignments.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * A task which describes what action should be performed for the specified
- * streaming computation ranges.
- */
+/// A task which describes what action should be performed for the specified
+/// streaming computation ranges.
class StreamingComputationTask {
- /** Contains ranges of a streaming computation this task should apply to. */
+ /// Contains ranges of a streaming computation this task should apply to.
core.List<StreamingComputationRanges> computationRanges;
- /** Describes the set of data disks this task should apply to. */
+
+ /// Describes the set of data disks this task should apply to.
core.List<MountedDataDisk> dataDisks;
- /**
- * A type of streaming computation task.
- * Possible string values are:
- * - "STREAMING_COMPUTATION_TASK_UNKNOWN" : The streaming computation task is
- * unknown, or unspecified.
- * - "STREAMING_COMPUTATION_TASK_STOP" : Stop processing specified streaming
- * computation range(s).
- * - "STREAMING_COMPUTATION_TASK_START" : Start processing specified streaming
- * computation range(s).
- */
+
+ /// A type of streaming computation task.
+ /// Possible string values are:
+ /// - "STREAMING_COMPUTATION_TASK_UNKNOWN" : The streaming computation task
+ /// is unknown, or unspecified.
+ /// - "STREAMING_COMPUTATION_TASK_STOP" : Stop processing specified streaming
+ /// computation range(s).
+ /// - "STREAMING_COMPUTATION_TASK_START" : Start processing specified
+ /// streaming computation range(s).
core.String taskType;
StreamingComputationTask();
StreamingComputationTask.fromJson(core.Map _json) {
if (_json.containsKey("computationRanges")) {
- computationRanges = _json["computationRanges"].map((value) => new StreamingComputationRanges.fromJson(value)).toList();
+ computationRanges = _json["computationRanges"]
+ .map((value) => new StreamingComputationRanges.fromJson(value))
+ .toList();
}
if (_json.containsKey("dataDisks")) {
- dataDisks = _json["dataDisks"].map((value) => new MountedDataDisk.fromJson(value)).toList();
+ dataDisks = _json["dataDisks"]
+ .map((value) => new MountedDataDisk.fromJson(value))
+ .toList();
}
if (_json.containsKey("taskType")) {
taskType = _json["taskType"];
@@ -7336,9 +7620,11 @@ class StreamingComputationTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (computationRanges != null) {
- _json["computationRanges"] = computationRanges.map((value) => (value).toJson()).toList();
+ _json["computationRanges"] =
+ computationRanges.map((value) => (value).toJson()).toList();
}
if (dataDisks != null) {
_json["dataDisks"] = dataDisks.map((value) => (value).toJson()).toList();
@@ -7350,32 +7636,32 @@ class StreamingComputationTask {
}
}
-/**
- * A task that carries configuration information for streaming computations.
- */
+/// A task that carries configuration information for streaming computations.
class StreamingConfigTask {
- /** Set of computation configuration information. */
+ /// Set of computation configuration information.
core.List<StreamingComputationConfig> streamingComputationConfigs;
- /** Map from user step names to state families. */
+
+ /// Map from user step names to state families.
core.Map<core.String, core.String> userStepToStateFamilyNameMap;
- /**
- * If present, the worker must use this endpoint to communicate with Windmill
- * Service dispatchers, otherwise the worker must continue to use whatever
- * endpoint it had been using.
- */
+
+ /// If present, the worker must use this endpoint to communicate with
+ /// Windmill
+ /// Service dispatchers, otherwise the worker must continue to use whatever
+ /// endpoint it had been using.
core.String windmillServiceEndpoint;
- /**
- * If present, the worker must use this port to communicate with Windmill
- * Service dispatchers. Only applicable when windmill_service_endpoint is
- * specified.
- */
+
+ /// If present, the worker must use this port to communicate with Windmill
+ /// Service dispatchers. Only applicable when windmill_service_endpoint is
+ /// specified.
core.String windmillServicePort;
StreamingConfigTask();
StreamingConfigTask.fromJson(core.Map _json) {
if (_json.containsKey("streamingComputationConfigs")) {
- streamingComputationConfigs = _json["streamingComputationConfigs"].map((value) => new StreamingComputationConfig.fromJson(value)).toList();
+ streamingComputationConfigs = _json["streamingComputationConfigs"]
+ .map((value) => new StreamingComputationConfig.fromJson(value))
+ .toList();
}
if (_json.containsKey("userStepToStateFamilyNameMap")) {
userStepToStateFamilyNameMap = _json["userStepToStateFamilyNameMap"];
@@ -7389,9 +7675,11 @@ class StreamingConfigTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (streamingComputationConfigs != null) {
- _json["streamingComputationConfigs"] = streamingComputationConfigs.map((value) => (value).toJson()).toList();
+ _json["streamingComputationConfigs"] =
+ streamingComputationConfigs.map((value) => (value).toJson()).toList();
}
if (userStepToStateFamilyNameMap != null) {
_json["userStepToStateFamilyNameMap"] = userStepToStateFamilyNameMap;
@@ -7406,21 +7694,20 @@ class StreamingConfigTask {
}
}
-/** A task which initializes part of a streaming Dataflow job. */
+/// A task which initializes part of a streaming Dataflow job.
class StreamingSetupTask {
- /** The user has requested drain. */
+ /// The user has requested drain.
core.bool drain;
- /**
- * The TCP port on which the worker should listen for messages from
- * other streaming computation workers.
- */
+
+ /// The TCP port on which the worker should listen for messages from
+ /// other streaming computation workers.
core.int receiveWorkPort;
- /** The global topology of the streaming Dataflow job. */
+
+ /// The global topology of the streaming Dataflow job.
TopologyConfig streamingComputationTopology;
- /**
- * The TCP port used by the worker to communicate with the Dataflow
- * worker harness.
- */
+
+ /// The TCP port used by the worker to communicate with the Dataflow
+ /// worker harness.
core.int workerHarnessPort;
StreamingSetupTask();
@@ -7433,7 +7720,8 @@ class StreamingSetupTask {
receiveWorkPort = _json["receiveWorkPort"];
}
if (_json.containsKey("streamingComputationTopology")) {
- streamingComputationTopology = new TopologyConfig.fromJson(_json["streamingComputationTopology"]);
+ streamingComputationTopology =
+ new TopologyConfig.fromJson(_json["streamingComputationTopology"]);
}
if (_json.containsKey("workerHarnessPort")) {
workerHarnessPort = _json["workerHarnessPort"];
@@ -7441,7 +7729,8 @@ class StreamingSetupTask {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (drain != null) {
_json["drain"] = drain;
}
@@ -7449,7 +7738,8 @@ class StreamingSetupTask {
_json["receiveWorkPort"] = receiveWorkPort;
}
if (streamingComputationTopology != null) {
- _json["streamingComputationTopology"] = (streamingComputationTopology).toJson();
+ _json["streamingComputationTopology"] =
+ (streamingComputationTopology).toJson();
}
if (workerHarnessPort != null) {
_json["workerHarnessPort"] = workerHarnessPort;
@@ -7458,13 +7748,12 @@ class StreamingSetupTask {
}
}
-/** Identifies the location of a streaming side input. */
+/// Identifies the location of a streaming side input.
class StreamingSideInputLocation {
- /** Identifies the state family where this side input is stored. */
+ /// Identifies the state family where this side input is stored.
core.String stateFamily;
- /**
- * Identifies the particular side input within the streaming Dataflow job.
- */
+
+ /// Identifies the particular side input within the streaming Dataflow job.
core.String tag;
StreamingSideInputLocation();
@@ -7479,7 +7768,8 @@ class StreamingSideInputLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (stateFamily != null) {
_json["stateFamily"] = stateFamily;
}
@@ -7490,15 +7780,11 @@ class StreamingSideInputLocation {
}
}
-/**
- * Identifies the location of a streaming computation stage, for
- * stage-to-stage communication.
- */
+/// Identifies the location of a streaming computation stage, for
+/// stage-to-stage communication.
class StreamingStageLocation {
- /**
- * Identifies the particular stream within the streaming Dataflow
- * job.
- */
+ /// Identifies the particular stream within the streaming Dataflow
+ /// job.
core.String streamId;
StreamingStageLocation();
@@ -7510,7 +7796,8 @@ class StreamingStageLocation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (streamId != null) {
_json["streamId"] = streamId;
}
@@ -7518,9 +7805,9 @@ class StreamingStageLocation {
}
}
-/** A metric value representing a list of strings. */
+/// A metric value representing a list of strings.
class StringList {
- /** Elements of the list. */
+ /// Elements of the list.
core.List<core.String> elements;
StringList();
@@ -7532,7 +7819,8 @@ class StringList {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (elements != null) {
_json["elements"] = elements;
}
@@ -7540,20 +7828,19 @@ class StringList {
}
}
-/**
- * A rich message format, including a human readable string, a key for
- * identifying the message, and structured data associated with the message for
- * programmatic consumption.
- */
+/// A rich message format, including a human readable string, a key for
+/// identifying the message, and structured data associated with the message
+/// for
+/// programmatic consumption.
class StructuredMessage {
- /**
- * Idenfier for this message type. Used by external systems to
- * internationalize or personalize message.
- */
+ /// Idenfier for this message type. Used by external systems to
+ /// internationalize or personalize message.
core.String messageKey;
- /** Human-readable version of message. */
+
+ /// Human-readable version of message.
core.String messageText;
- /** The structured data associated with this message. */
+
+ /// The structured data associated with this message.
core.List<Parameter> parameters;
StructuredMessage();
@@ -7566,12 +7853,15 @@ class StructuredMessage {
messageText = _json["messageText"];
}
if (_json.containsKey("parameters")) {
- parameters = _json["parameters"].map((value) => new Parameter.fromJson(value)).toList();
+ parameters = _json["parameters"]
+ .map((value) => new Parameter.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (messageKey != null) {
_json["messageKey"] = messageKey;
}
@@ -7579,91 +7869,96 @@ class StructuredMessage {
_json["messageText"] = messageText;
}
if (parameters != null) {
- _json["parameters"] = parameters.map((value) => (value).toJson()).toList();
+ _json["parameters"] =
+ parameters.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** Taskrunner configuration settings. */
+/// Taskrunner configuration settings.
class TaskRunnerSettings {
- /** Whether to also send taskrunner log info to stderr. */
+ /// Whether to also send taskrunner log info to stderr.
core.bool alsologtostderr;
- /** The location on the worker for task-specific subdirectories. */
+
+ /// The location on the worker for task-specific subdirectories.
core.String baseTaskDir;
- /**
- * The base URL for the taskrunner to use when accessing Google Cloud APIs.
- *
- * When workers access Google Cloud APIs, they logically do so via
- * relative URLs. If this field is specified, it supplies the base
- * URL to use for resolving these relative URLs. The normative
- * algorithm used is defined by RFC 1808, "Relative Uniform Resource
- * Locators".
- *
- * If not specified, the default value is "http://www.googleapis.com/"
- */
+
+ /// The base URL for the taskrunner to use when accessing Google Cloud APIs.
+ ///
+ /// When workers access Google Cloud APIs, they logically do so via
+ /// relative URLs. If this field is specified, it supplies the base
+ /// URL to use for resolving these relative URLs. The normative
+ /// algorithm used is defined by RFC 1808, "Relative Uniform Resource
+ /// Locators".
+ ///
+ /// If not specified, the default value is "http://www.googleapis.com/"
core.String baseUrl;
- /** The file to store preprocessing commands in. */
+
+ /// The file to store preprocessing commands in.
core.String commandlinesFileName;
- /** Whether to continue taskrunner if an exception is hit. */
+
+ /// Whether to continue taskrunner if an exception is hit.
core.bool continueOnException;
- /** The API version of endpoint, e.g. "v1b3" */
+
+ /// The API version of endpoint, e.g. "v1b3"
core.String dataflowApiVersion;
- /** The command to launch the worker harness. */
+
+ /// The command to launch the worker harness.
core.String harnessCommand;
- /** The suggested backend language. */
+
+ /// The suggested backend language.
core.String languageHint;
- /** The directory on the VM to store logs. */
+
+ /// The directory on the VM to store logs.
core.String logDir;
- /**
- * Whether to send taskrunner log info to Google Compute Engine VM serial
- * console.
- */
+
+ /// Whether to send taskrunner log info to Google Compute Engine VM serial
+ /// console.
core.bool logToSerialconsole;
- /**
- * Indicates where to put logs. If this is not specified, the logs
- * will not be uploaded.
- *
- * The supported resource type is:
- *
- * Google Cloud Storage:
- * storage.googleapis.com/{bucket}/{object}
- * bucket.storage.googleapis.com/{object}
- */
+
+ /// Indicates where to put logs. If this is not specified, the logs
+ /// will not be uploaded.
+ ///
+ /// The supported resource type is:
+ ///
+ /// Google Cloud Storage:
+ /// storage.googleapis.com/{bucket}/{object}
+ /// bucket.storage.googleapis.com/{object}
core.String logUploadLocation;
- /**
- * The OAuth2 scopes to be requested by the taskrunner in order to
- * access the Cloud Dataflow API.
- */
+
+ /// The OAuth2 scopes to be requested by the taskrunner in order to
+ /// access the Cloud Dataflow API.
core.List<core.String> oauthScopes;
- /** The settings to pass to the parallel worker harness. */
+
+ /// The settings to pass to the parallel worker harness.
WorkerSettings parallelWorkerSettings;
- /** The streaming worker main class name. */
+
+ /// The streaming worker main class name.
core.String streamingWorkerMainClass;
- /**
- * The UNIX group ID on the worker VM to use for tasks launched by
- * taskrunner; e.g. "wheel".
- */
+
+ /// The UNIX group ID on the worker VM to use for tasks launched by
+ /// taskrunner; e.g. "wheel".
core.String taskGroup;
- /**
- * The UNIX user ID on the worker VM to use for tasks launched by
- * taskrunner; e.g. "root".
- */
+
+ /// The UNIX user ID on the worker VM to use for tasks launched by
+ /// taskrunner; e.g. "root".
core.String taskUser;
- /**
- * The prefix of the resources the taskrunner should use for
- * temporary storage.
- *
- * The supported resource type is:
- *
- * Google Cloud Storage:
- * storage.googleapis.com/{bucket}/{object}
- * bucket.storage.googleapis.com/{object}
- */
+
+ /// The prefix of the resources the taskrunner should use for
+ /// temporary storage.
+ ///
+ /// The supported resource type is:
+ ///
+ /// Google Cloud Storage:
+ /// storage.googleapis.com/{bucket}/{object}
+ /// bucket.storage.googleapis.com/{object}
core.String tempStoragePrefix;
- /** The ID string of the VM. */
+
+ /// The ID string of the VM.
core.String vmId;
- /** The file to store the workflow in. */
+
+ /// The file to store the workflow in.
core.String workflowFileName;
TaskRunnerSettings();
@@ -7706,7 +8001,8 @@ class TaskRunnerSettings {
oauthScopes = _json["oauthScopes"];
}
if (_json.containsKey("parallelWorkerSettings")) {
- parallelWorkerSettings = new WorkerSettings.fromJson(_json["parallelWorkerSettings"]);
+ parallelWorkerSettings =
+ new WorkerSettings.fromJson(_json["parallelWorkerSettings"]);
}
if (_json.containsKey("streamingWorkerMainClass")) {
streamingWorkerMainClass = _json["streamingWorkerMainClass"];
@@ -7729,7 +8025,8 @@ class TaskRunnerSettings {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (alsologtostderr != null) {
_json["alsologtostderr"] = alsologtostderr;
}
@@ -7791,13 +8088,15 @@ class TaskRunnerSettings {
}
}
-/** Metadata describing a template. */
+/// Metadata describing a template.
class TemplateMetadata {
- /** Optional. A description of the template. */
+ /// Optional. A description of the template.
core.String description;
- /** Required. The name of the template. */
+
+ /// Required. The name of the template.
core.String name;
- /** The parameters for the template. */
+
+ /// The parameters for the template.
core.List<ParameterMetadata> parameters;
TemplateMetadata();
@@ -7810,12 +8109,15 @@ class TemplateMetadata {
name = _json["name"];
}
if (_json.containsKey("parameters")) {
- parameters = _json["parameters"].map((value) => new ParameterMetadata.fromJson(value)).toList();
+ parameters = _json["parameters"]
+ .map((value) => new ParameterMetadata.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (description != null) {
_json["description"] = description;
}
@@ -7823,36 +8125,43 @@ class TemplateMetadata {
_json["name"] = name;
}
if (parameters != null) {
- _json["parameters"] = parameters.map((value) => (value).toJson()).toList();
+ _json["parameters"] =
+ parameters.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * Global topology of the streaming Dataflow job, including all
- * computations and their sharded locations.
- */
+/// Global topology of the streaming Dataflow job, including all
+/// computations and their sharded locations.
class TopologyConfig {
- /** The computations associated with a streaming Dataflow job. */
+ /// The computations associated with a streaming Dataflow job.
core.List<ComputationTopology> computations;
- /** The disks assigned to a streaming Dataflow job. */
+
+ /// The disks assigned to a streaming Dataflow job.
core.List<DataDiskAssignment> dataDiskAssignments;
- /** The size (in bits) of keys that will be assigned to source messages. */
+
+ /// The size (in bits) of keys that will be assigned to source messages.
core.int forwardingKeyBits;
- /** Version number for persistent state. */
+
+ /// Version number for persistent state.
core.int persistentStateVersion;
- /** Maps user stage names to stable computation names. */
+
+ /// Maps user stage names to stable computation names.
core.Map<core.String, core.String> userStageToComputationNameMap;
TopologyConfig();
TopologyConfig.fromJson(core.Map _json) {
if (_json.containsKey("computations")) {
- computations = _json["computations"].map((value) => new ComputationTopology.fromJson(value)).toList();
+ computations = _json["computations"]
+ .map((value) => new ComputationTopology.fromJson(value))
+ .toList();
}
if (_json.containsKey("dataDiskAssignments")) {
- dataDiskAssignments = _json["dataDiskAssignments"].map((value) => new DataDiskAssignment.fromJson(value)).toList();
+ dataDiskAssignments = _json["dataDiskAssignments"]
+ .map((value) => new DataDiskAssignment.fromJson(value))
+ .toList();
}
if (_json.containsKey("forwardingKeyBits")) {
forwardingKeyBits = _json["forwardingKeyBits"];
@@ -7866,12 +8175,15 @@ class TopologyConfig {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (computations != null) {
- _json["computations"] = computations.map((value) => (value).toJson()).toList();
+ _json["computations"] =
+ computations.map((value) => (value).toJson()).toList();
}
if (dataDiskAssignments != null) {
- _json["dataDiskAssignments"] = dataDiskAssignments.map((value) => (value).toJson()).toList();
+ _json["dataDiskAssignments"] =
+ dataDiskAssignments.map((value) => (value).toJson()).toList();
}
if (forwardingKeyBits != null) {
_json["forwardingKeyBits"] = forwardingKeyBits;
@@ -7886,40 +8198,45 @@ class TopologyConfig {
}
}
-/** Description of the type, names/ids, and input/outputs for a transform. */
+/// Description of the type, names/ids, and input/outputs for a transform.
class TransformSummary {
- /** Transform-specific display data. */
+ /// Transform-specific display data.
core.List<DisplayData> displayData;
- /** SDK generated id of this transform instance. */
+
+ /// SDK generated id of this transform instance.
core.String id;
- /** User names for all collection inputs to this transform. */
+
+ /// User names for all collection inputs to this transform.
core.List<core.String> inputCollectionName;
- /**
- * Type of transform.
- * Possible string values are:
- * - "UNKNOWN_KIND" : Unrecognized transform type.
- * - "PAR_DO_KIND" : ParDo transform.
- * - "GROUP_BY_KEY_KIND" : Group By Key transform.
- * - "FLATTEN_KIND" : Flatten transform.
- * - "READ_KIND" : Read transform.
- * - "WRITE_KIND" : Write transform.
- * - "CONSTANT_KIND" : Constructs from a constant value, such as with
- * Create.of.
- * - "SINGLETON_KIND" : Creates a Singleton view of a collection.
- * - "SHUFFLE_KIND" : Opening or closing a shuffle session, often as part of a
- * GroupByKey.
- */
+
+ /// Type of transform.
+ /// Possible string values are:
+ /// - "UNKNOWN_KIND" : Unrecognized transform type.
+ /// - "PAR_DO_KIND" : ParDo transform.
+ /// - "GROUP_BY_KEY_KIND" : Group By Key transform.
+ /// - "FLATTEN_KIND" : Flatten transform.
+ /// - "READ_KIND" : Read transform.
+ /// - "WRITE_KIND" : Write transform.
+ /// - "CONSTANT_KIND" : Constructs from a constant value, such as with
+ /// Create.of.
+ /// - "SINGLETON_KIND" : Creates a Singleton view of a collection.
+ /// - "SHUFFLE_KIND" : Opening or closing a shuffle session, often as part of
+ /// a GroupByKey.
core.String kind;
- /** User provided name for this transform instance. */
+
+ /// User provided name for this transform instance.
core.String name;
- /** User names for all collection outputs to this transform. */
+
+ /// User names for all collection outputs to this transform.
core.List<core.String> outputCollectionName;
TransformSummary();
TransformSummary.fromJson(core.Map _json) {
if (_json.containsKey("displayData")) {
- displayData = _json["displayData"].map((value) => new DisplayData.fromJson(value)).toList();
+ displayData = _json["displayData"]
+ .map((value) => new DisplayData.fromJson(value))
+ .toList();
}
if (_json.containsKey("id")) {
id = _json["id"];
@@ -7939,9 +8256,11 @@ class TransformSummary {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (displayData != null) {
- _json["displayData"] = displayData.map((value) => (value).toJson()).toList();
+ _json["displayData"] =
+ displayData.map((value) => (value).toJson()).toList();
}
if (id != null) {
_json["id"] = id;
@@ -7962,43 +8281,53 @@ class TransformSummary {
}
}
-/**
- * WorkItem represents basic information about a WorkItem to be executed
- * in the cloud.
- */
+/// WorkItem represents basic information about a WorkItem to be executed
+/// in the cloud.
class WorkItem {
- /** Work item-specific configuration as an opaque blob. */
+ /// Work item-specific configuration as an opaque blob.
core.String configuration;
- /** Identifies this WorkItem. */
+
+ /// Identifies this WorkItem.
core.String id;
- /** The initial index to use when reporting the status of the WorkItem. */
+
+ /// The initial index to use when reporting the status of the WorkItem.
core.String initialReportIndex;
- /** Identifies the workflow job this WorkItem belongs to. */
+
+ /// Identifies the workflow job this WorkItem belongs to.
core.String jobId;
- /** Time when the lease on this Work will expire. */
+
+ /// Time when the lease on this Work will expire.
core.String leaseExpireTime;
- /** Additional information for MapTask WorkItems. */
+
+ /// Additional information for MapTask WorkItems.
MapTask mapTask;
- /**
- * Any required packages that need to be fetched in order to execute
- * this WorkItem.
- */
+
+ /// Any required packages that need to be fetched in order to execute
+ /// this WorkItem.
core.List<Package> packages;
- /** Identifies the cloud project this WorkItem belongs to. */
+
+ /// Identifies the cloud project this WorkItem belongs to.
core.String projectId;
- /** Recommended reporting interval. */
+
+ /// Recommended reporting interval.
core.String reportStatusInterval;
- /** Additional information for SeqMapTask WorkItems. */
+
+ /// Additional information for SeqMapTask WorkItems.
SeqMapTask seqMapTask;
- /** Additional information for ShellTask WorkItems. */
+
+ /// Additional information for ShellTask WorkItems.
ShellTask shellTask;
- /** Additional information for source operation WorkItems. */
+
+ /// Additional information for source operation WorkItems.
SourceOperationRequest sourceOperationTask;
- /** Additional information for StreamingComputationTask WorkItems. */
+
+ /// Additional information for StreamingComputationTask WorkItems.
StreamingComputationTask streamingComputationTask;
- /** Additional information for StreamingConfigTask WorkItems. */
+
+ /// Additional information for StreamingConfigTask WorkItems.
StreamingConfigTask streamingConfigTask;
- /** Additional information for StreamingSetupTask WorkItems. */
+
+ /// Additional information for StreamingSetupTask WorkItems.
StreamingSetupTask streamingSetupTask;
WorkItem();
@@ -8023,7 +8352,9 @@ class WorkItem {
mapTask = new MapTask.fromJson(_json["mapTask"]);
}
if (_json.containsKey("packages")) {
- packages = _json["packages"].map((value) => new Package.fromJson(value)).toList();
+ packages = _json["packages"]
+ .map((value) => new Package.fromJson(value))
+ .toList();
}
if (_json.containsKey("projectId")) {
projectId = _json["projectId"];
@@ -8038,21 +8369,26 @@ class WorkItem {
shellTask = new ShellTask.fromJson(_json["shellTask"]);
}
if (_json.containsKey("sourceOperationTask")) {
- sourceOperationTask = new SourceOperationRequest.fromJson(_json["sourceOperationTask"]);
+ sourceOperationTask =
+ new SourceOperationRequest.fromJson(_json["sourceOperationTask"]);
}
if (_json.containsKey("streamingComputationTask")) {
- streamingComputationTask = new StreamingComputationTask.fromJson(_json["streamingComputationTask"]);
+ streamingComputationTask = new StreamingComputationTask.fromJson(
+ _json["streamingComputationTask"]);
}
if (_json.containsKey("streamingConfigTask")) {
- streamingConfigTask = new StreamingConfigTask.fromJson(_json["streamingConfigTask"]);
+ streamingConfigTask =
+ new StreamingConfigTask.fromJson(_json["streamingConfigTask"]);
}
if (_json.containsKey("streamingSetupTask")) {
- streamingSetupTask = new StreamingSetupTask.fromJson(_json["streamingSetupTask"]);
+ streamingSetupTask =
+ new StreamingSetupTask.fromJson(_json["streamingSetupTask"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (configuration != null) {
_json["configuration"] = configuration;
}
@@ -8102,46 +8438,43 @@ class WorkItem {
}
}
-/**
- * The Dataflow service's idea of the current state of a WorkItem
- * being processed by a worker.
- */
+/// The Dataflow service's idea of the current state of a WorkItem
+/// being processed by a worker.
class WorkItemServiceState {
- /**
- * Other data returned by the service, specific to the particular
- * worker harness.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// Other data returned by the service, specific to the particular
+ /// worker harness.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> harnessData;
- /** Time at which the current lease will expire. */
+
+ /// Time at which the current lease will expire.
core.String leaseExpireTime;
- /**
- * The short ids that workers should use in subsequent metric updates.
- * Workers should strive to use short ids whenever possible, but it is ok
- * to request the short_id again if a worker lost track of it
- * (e.g. if the worker is recovering from a crash).
- * NOTE: it is possible that the response may have short ids for a subset
- * of the metrics.
- */
+
+ /// The short ids that workers should use in subsequent metric updates.
+ /// Workers should strive to use short ids whenever possible, but it is ok
+ /// to request the short_id again if a worker lost track of it
+ /// (e.g. if the worker is recovering from a crash).
+ /// NOTE: it is possible that the response may have short ids for a subset
+ /// of the metrics.
core.List<MetricShortId> metricShortId;
- /**
- * The index value to use for the next report sent by the worker.
- * Note: If the report call fails for whatever reason, the worker should
- * reuse this index for subsequent report attempts.
- */
+
+ /// The index value to use for the next report sent by the worker.
+ /// Note: If the report call fails for whatever reason, the worker should
+ /// reuse this index for subsequent report attempts.
core.String nextReportIndex;
- /** New recommended reporting interval. */
+
+ /// New recommended reporting interval.
core.String reportStatusInterval;
- /**
- * The progress point in the WorkItem where the Dataflow service
- * suggests that the worker truncate the task.
- */
+
+ /// The progress point in the WorkItem where the Dataflow service
+ /// suggests that the worker truncate the task.
ApproximateSplitRequest splitRequest;
- /** DEPRECATED in favor of split_request. */
+
+ /// DEPRECATED in favor of split_request.
ApproximateProgress suggestedStopPoint;
- /** Obsolete, always empty. */
+
+ /// Obsolete, always empty.
Position suggestedStopPosition;
WorkItemServiceState();
@@ -8154,7 +8487,9 @@ class WorkItemServiceState {
leaseExpireTime = _json["leaseExpireTime"];
}
if (_json.containsKey("metricShortId")) {
- metricShortId = _json["metricShortId"].map((value) => new MetricShortId.fromJson(value)).toList();
+ metricShortId = _json["metricShortId"]
+ .map((value) => new MetricShortId.fromJson(value))
+ .toList();
}
if (_json.containsKey("nextReportIndex")) {
nextReportIndex = _json["nextReportIndex"];
@@ -8163,18 +8498,22 @@ class WorkItemServiceState {
reportStatusInterval = _json["reportStatusInterval"];
}
if (_json.containsKey("splitRequest")) {
- splitRequest = new ApproximateSplitRequest.fromJson(_json["splitRequest"]);
+ splitRequest =
+ new ApproximateSplitRequest.fromJson(_json["splitRequest"]);
}
if (_json.containsKey("suggestedStopPoint")) {
- suggestedStopPoint = new ApproximateProgress.fromJson(_json["suggestedStopPoint"]);
+ suggestedStopPoint =
+ new ApproximateProgress.fromJson(_json["suggestedStopPoint"]);
}
if (_json.containsKey("suggestedStopPosition")) {
- suggestedStopPosition = new Position.fromJson(_json["suggestedStopPosition"]);
+ suggestedStopPosition =
+ new Position.fromJson(_json["suggestedStopPosition"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (harnessData != null) {
_json["harnessData"] = harnessData;
}
@@ -8182,7 +8521,8 @@ class WorkItemServiceState {
_json["leaseExpireTime"] = leaseExpireTime;
}
if (metricShortId != null) {
- _json["metricShortId"] = metricShortId.map((value) => (value).toJson()).toList();
+ _json["metricShortId"] =
+ metricShortId.map((value) => (value).toJson()).toList();
}
if (nextReportIndex != null) {
_json["nextReportIndex"] = nextReportIndex;
@@ -8203,84 +8543,88 @@ class WorkItemServiceState {
}
}
-/** Conveys a worker's progress through the work described by a WorkItem. */
+/// Conveys a worker's progress through the work described by a WorkItem.
class WorkItemStatus {
- /** True if the WorkItem was completed (successfully or unsuccessfully). */
+ /// True if the WorkItem was completed (successfully or unsuccessfully).
core.bool completed;
- /** Worker output counters for this WorkItem. */
+
+ /// Worker output counters for this WorkItem.
core.List<CounterUpdate> counterUpdates;
- /** See documentation of stop_position. */
+
+ /// See documentation of stop_position.
DynamicSourceSplit dynamicSourceSplit;
- /**
- * Specifies errors which occurred during processing. If errors are
- * provided, and completed = true, then the WorkItem is considered
- * to have failed.
- */
+
+ /// Specifies errors which occurred during processing. If errors are
+ /// provided, and completed = true, then the WorkItem is considered
+ /// to have failed.
core.List<Status> errors;
- /** DEPRECATED in favor of counter_updates. */
+
+ /// DEPRECATED in favor of counter_updates.
core.List<MetricUpdate> metricUpdates;
- /** DEPRECATED in favor of reported_progress. */
+
+ /// DEPRECATED in favor of reported_progress.
ApproximateProgress progress;
- /**
- * The report index. When a WorkItem is leased, the lease will
- * contain an initial report index. When a WorkItem's status is
- * reported to the system, the report should be sent with
- * that report index, and the response will contain the index the
- * worker should use for the next report. Reports received with
- * unexpected index values will be rejected by the service.
- *
- * In order to preserve idempotency, the worker should not alter the
- * contents of a report, even if the worker must submit the same
- * report multiple times before getting back a response. The worker
- * should not submit a subsequent report until the response for the
- * previous report had been received from the service.
- */
+
+ /// The report index. When a WorkItem is leased, the lease will
+ /// contain an initial report index. When a WorkItem's status is
+ /// reported to the system, the report should be sent with
+ /// that report index, and the response will contain the index the
+ /// worker should use for the next report. Reports received with
+ /// unexpected index values will be rejected by the service.
+ ///
+ /// In order to preserve idempotency, the worker should not alter the
+ /// contents of a report, even if the worker must submit the same
+ /// report multiple times before getting back a response. The worker
+ /// should not submit a subsequent report until the response for the
+ /// previous report had been received from the service.
core.String reportIndex;
- /** The worker's progress through this WorkItem. */
+
+ /// The worker's progress through this WorkItem.
ApproximateReportedProgress reportedProgress;
- /** Amount of time the worker requests for its lease. */
+
+ /// Amount of time the worker requests for its lease.
core.String requestedLeaseDuration;
- /** DEPRECATED in favor of dynamic_source_split. */
+
+ /// DEPRECATED in favor of dynamic_source_split.
SourceFork sourceFork;
- /**
- * If the work item represented a SourceOperationRequest, and the work
- * is completed, contains the result of the operation.
- */
+
+ /// If the work item represented a SourceOperationRequest, and the work
+ /// is completed, contains the result of the operation.
SourceOperationResponse sourceOperationResponse;
- /**
- * A worker may split an active map task in two parts, "primary" and
- * "residual", continuing to process the primary part and returning the
- * residual part into the pool of available work.
- * This event is called a "dynamic split" and is critical to the dynamic
- * work rebalancing feature. The two obtained sub-tasks are called
- * "parts" of the split.
- * The parts, if concatenated, must represent the same input as would
- * be read by the current task if the split did not happen.
- * The exact way in which the original task is decomposed into the two
- * parts is specified either as a position demarcating them
- * (stop_position), or explicitly as two DerivedSources, if this
- * task consumes a user-defined source type (dynamic_source_split).
- *
- * The "current" task is adjusted as a result of the split: after a task
- * with range [A, B) sends a stop_position update at C, its range is
- * considered to be [A, C), e.g.:
- * * Progress should be interpreted relative to the new range, e.g.
- * "75% completed" means "75% of [A, C) completed"
- * * The worker should interpret proposed_stop_position relative to the
- * new range, e.g. "split at 68%" should be interpreted as
- * "split at 68% of [A, C)".
- * * If the worker chooses to split again using stop_position, only
- * stop_positions in [A, C) will be accepted.
- * * Etc.
- * dynamic_source_split has similar semantics: e.g., if a task with
- * source S splits using dynamic_source_split into {P, R}
- * (where P and R must be together equivalent to S), then subsequent
- * progress and proposed_stop_position should be interpreted relative
- * to P, and in a potential subsequent dynamic_source_split into {P', R'},
- * P' and R' must be together equivalent to P, etc.
- */
+
+ /// A worker may split an active map task in two parts, "primary" and
+ /// "residual", continuing to process the primary part and returning the
+ /// residual part into the pool of available work.
+ /// This event is called a "dynamic split" and is critical to the dynamic
+ /// work rebalancing feature. The two obtained sub-tasks are called
+ /// "parts" of the split.
+ /// The parts, if concatenated, must represent the same input as would
+ /// be read by the current task if the split did not happen.
+ /// The exact way in which the original task is decomposed into the two
+ /// parts is specified either as a position demarcating them
+ /// (stop_position), or explicitly as two DerivedSources, if this
+ /// task consumes a user-defined source type (dynamic_source_split).
+ ///
+ /// The "current" task is adjusted as a result of the split: after a task
+ /// with range [A, B) sends a stop_position update at C, its range is
+ /// considered to be [A, C), e.g.:
+ /// * Progress should be interpreted relative to the new range, e.g.
+ /// "75% completed" means "75% of [A, C) completed"
+ /// * The worker should interpret proposed_stop_position relative to the
+ /// new range, e.g. "split at 68%" should be interpreted as
+ /// "split at 68% of [A, C)".
+ /// * If the worker chooses to split again using stop_position, only
+ /// stop_positions in [A, C) will be accepted.
+ /// * Etc.
+ /// dynamic_source_split has similar semantics: e.g., if a task with
+ /// source S splits using dynamic_source_split into {P, R}
+ /// (where P and R must be together equivalent to S), then subsequent
+ /// progress and proposed_stop_position should be interpreted relative
+ /// to P, and in a potential subsequent dynamic_source_split into {P', R'},
+ /// P' and R' must be together equivalent to P, etc.
Position stopPosition;
- /** Identifies the WorkItem. */
+
+ /// Identifies the WorkItem.
core.String workItemId;
WorkItemStatus();
@@ -8290,16 +8634,22 @@ class WorkItemStatus {
completed = _json["completed"];
}
if (_json.containsKey("counterUpdates")) {
- counterUpdates = _json["counterUpdates"].map((value) => new CounterUpdate.fromJson(value)).toList();
+ counterUpdates = _json["counterUpdates"]
+ .map((value) => new CounterUpdate.fromJson(value))
+ .toList();
}
if (_json.containsKey("dynamicSourceSplit")) {
- dynamicSourceSplit = new DynamicSourceSplit.fromJson(_json["dynamicSourceSplit"]);
+ dynamicSourceSplit =
+ new DynamicSourceSplit.fromJson(_json["dynamicSourceSplit"]);
}
if (_json.containsKey("errors")) {
- errors = _json["errors"].map((value) => new Status.fromJson(value)).toList();
+ errors =
+ _json["errors"].map((value) => new Status.fromJson(value)).toList();
}
if (_json.containsKey("metricUpdates")) {
- metricUpdates = _json["metricUpdates"].map((value) => new MetricUpdate.fromJson(value)).toList();
+ metricUpdates = _json["metricUpdates"]
+ .map((value) => new MetricUpdate.fromJson(value))
+ .toList();
}
if (_json.containsKey("progress")) {
progress = new ApproximateProgress.fromJson(_json["progress"]);
@@ -8308,7 +8658,8 @@ class WorkItemStatus {
reportIndex = _json["reportIndex"];
}
if (_json.containsKey("reportedProgress")) {
- reportedProgress = new ApproximateReportedProgress.fromJson(_json["reportedProgress"]);
+ reportedProgress =
+ new ApproximateReportedProgress.fromJson(_json["reportedProgress"]);
}
if (_json.containsKey("requestedLeaseDuration")) {
requestedLeaseDuration = _json["requestedLeaseDuration"];
@@ -8317,7 +8668,8 @@ class WorkItemStatus {
sourceFork = new SourceFork.fromJson(_json["sourceFork"]);
}
if (_json.containsKey("sourceOperationResponse")) {
- sourceOperationResponse = new SourceOperationResponse.fromJson(_json["sourceOperationResponse"]);
+ sourceOperationResponse = new SourceOperationResponse.fromJson(
+ _json["sourceOperationResponse"]);
}
if (_json.containsKey("stopPosition")) {
stopPosition = new Position.fromJson(_json["stopPosition"]);
@@ -8328,12 +8680,14 @@ class WorkItemStatus {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (completed != null) {
_json["completed"] = completed;
}
if (counterUpdates != null) {
- _json["counterUpdates"] = counterUpdates.map((value) => (value).toJson()).toList();
+ _json["counterUpdates"] =
+ counterUpdates.map((value) => (value).toJson()).toList();
}
if (dynamicSourceSplit != null) {
_json["dynamicSourceSplit"] = (dynamicSourceSplit).toJson();
@@ -8342,7 +8696,8 @@ class WorkItemStatus {
_json["errors"] = errors.map((value) => (value).toJson()).toList();
}
if (metricUpdates != null) {
- _json["metricUpdates"] = metricUpdates.map((value) => (value).toJson()).toList();
+ _json["metricUpdates"] =
+ metricUpdates.map((value) => (value).toJson()).toList();
}
if (progress != null) {
_json["progress"] = (progress).toJson();
@@ -8372,33 +8727,31 @@ class WorkItemStatus {
}
}
-/**
- * WorkerHealthReport contains information about the health of a worker.
- *
- * The VM should be identified by the labels attached to the WorkerMessage that
- * this health ping belongs to.
- */
+/// WorkerHealthReport contains information about the health of a worker.
+///
+/// The VM should be identified by the labels attached to the WorkerMessage
+/// that
+/// this health ping belongs to.
class WorkerHealthReport {
- /**
- * The pods running on the worker. See:
- * http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html#_v1_pod
- *
- * This field is used by the worker to send the status of the indvidual
- * containers running on each worker.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+ /// The pods running on the worker. See:
+ /// http://kubernetes.io/v1.1/docs/api-reference/v1/definitions.html#_v1_pod
+ ///
+ /// This field is used by the worker to send the status of the indvidual
+ /// containers running on each worker.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.List<core.Map<core.String, core.Object>> pods;
- /**
- * The interval at which the worker is sending health reports.
- * The default value of 0 should be interpreted as the field is not being
- * explicitly set by the worker.
- */
+
+ /// The interval at which the worker is sending health reports.
+ /// The default value of 0 should be interpreted as the field is not being
+ /// explicitly set by the worker.
core.String reportInterval;
- /** Whether the VM is healthy. */
+
+ /// Whether the VM is healthy.
core.bool vmIsHealthy;
- /** The time the VM was booted. */
+
+ /// The time the VM was booted.
core.String vmStartupTime;
WorkerHealthReport();
@@ -8419,7 +8772,8 @@ class WorkerHealthReport {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (pods != null) {
_json["pods"] = pods;
}
@@ -8436,18 +8790,15 @@ class WorkerHealthReport {
}
}
-/**
- * WorkerHealthReportResponse contains information returned to the worker
- * in response to a health ping.
- */
+/// WorkerHealthReportResponse contains information returned to the worker
+/// in response to a health ping.
class WorkerHealthReportResponse {
- /**
- * A positive value indicates the worker should change its reporting interval
- * to the specified value.
- *
- * The default value of zero means no change in report rate is requested by
- * the server.
- */
+ /// A positive value indicates the worker should change its reporting
+ /// interval
+ /// to the specified value.
+ ///
+ /// The default value of zero means no change in report rate is requested by
+ /// the server.
core.String reportInterval;
WorkerHealthReportResponse();
@@ -8459,7 +8810,8 @@ class WorkerHealthReportResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (reportInterval != null) {
_json["reportInterval"] = reportInterval;
}
@@ -8467,30 +8819,36 @@ class WorkerHealthReportResponse {
}
}
-/** WorkerMessage provides information to the backend about a worker. */
+/// WorkerMessage provides information to the backend about a worker.
class WorkerMessage {
- /**
- * Labels are used to group WorkerMessages.
- * For example, a worker_message about a particular container
- * might have the labels:
- * { "JOB_ID": "2015-04-22",
- * "WORKER_ID": "wordcount-vm-2015…"
- * "CONTAINER_TYPE": "worker",
- * "CONTAINER_ID": "ac1234def"}
- * Label tags typically correspond to Label enum values. However, for ease
- * of development other strings can be used as tags. LABEL_UNSPECIFIED should
- * not be used here.
- */
+ /// Labels are used to group WorkerMessages.
+ /// For example, a worker_message about a particular container
+ /// might have the labels:
+ /// { "JOB_ID": "2015-04-22",
+ /// "WORKER_ID": "wordcount-vm-2015…"
+ /// "CONTAINER_TYPE": "worker",
+ /// "CONTAINER_ID": "ac1234def"}
+ /// Label tags typically correspond to Label enum values. However, for ease
+ /// of development other strings can be used as tags. LABEL_UNSPECIFIED
+ /// should
+ /// not be used here.
core.Map<core.String, core.String> labels;
- /** The timestamp of the worker_message. */
+
+ /// The timestamp of the worker_message.
core.String time;
- /** The health of a worker. */
+
+ /// The health of a worker.
WorkerHealthReport workerHealthReport;
- /** A worker message code. */
+
+ /// A worker message code.
WorkerMessageCode workerMessageCode;
- /** Resource metrics reported by workers. */
+
+ /// Resource metrics reported by workers.
ResourceUtilizationReport workerMetrics;
+ /// Shutdown notice by workers.
+ WorkerShutdownNotice workerShutdownNotice;
+
WorkerMessage();
WorkerMessage.fromJson(core.Map _json) {
@@ -8501,18 +8859,26 @@ class WorkerMessage {
time = _json["time"];
}
if (_json.containsKey("workerHealthReport")) {
- workerHealthReport = new WorkerHealthReport.fromJson(_json["workerHealthReport"]);
+ workerHealthReport =
+ new WorkerHealthReport.fromJson(_json["workerHealthReport"]);
}
if (_json.containsKey("workerMessageCode")) {
- workerMessageCode = new WorkerMessageCode.fromJson(_json["workerMessageCode"]);
+ workerMessageCode =
+ new WorkerMessageCode.fromJson(_json["workerMessageCode"]);
}
if (_json.containsKey("workerMetrics")) {
- workerMetrics = new ResourceUtilizationReport.fromJson(_json["workerMetrics"]);
+ workerMetrics =
+ new ResourceUtilizationReport.fromJson(_json["workerMetrics"]);
+ }
+ if (_json.containsKey("workerShutdownNotice")) {
+ workerShutdownNotice =
+ new WorkerShutdownNotice.fromJson(_json["workerShutdownNotice"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (labels != null) {
_json["labels"] = labels;
}
@@ -8528,60 +8894,60 @@ class WorkerMessage {
if (workerMetrics != null) {
_json["workerMetrics"] = (workerMetrics).toJson();
}
+ if (workerShutdownNotice != null) {
+ _json["workerShutdownNotice"] = (workerShutdownNotice).toJson();
+ }
return _json;
}
}
-/**
- * A message code is used to report status and error messages to the service.
- * The message codes are intended to be machine readable. The service will
- * take care of translating these into user understandable messages if
- * necessary.
- *
- * Example use cases:
- * 1. Worker processes reporting successful startup.
- * 2. Worker processes reporting specific errors (e.g. package staging
- * failure).
- */
+/// A message code is used to report status and error messages to the service.
+/// The message codes are intended to be machine readable. The service will
+/// take care of translating these into user understandable messages if
+/// necessary.
+///
+/// Example use cases:
+/// 1. Worker processes reporting successful startup.
+/// 2. Worker processes reporting specific errors (e.g. package staging
+/// failure).
class WorkerMessageCode {
- /**
- * The code is a string intended for consumption by a machine that identifies
- * the type of message being sent.
- * Examples:
- * 1. "HARNESS_STARTED" might be used to indicate the worker harness has
- * started.
- * 2. "GCS_DOWNLOAD_ERROR" might be used to indicate an error downloading
- * a GCS file as part of the boot process of one of the worker containers.
- *
- * This is a string and not an enum to make it easy to add new codes without
- * waiting for an API change.
- */
+ /// The code is a string intended for consumption by a machine that
+ /// identifies
+ /// the type of message being sent.
+ /// Examples:
+ /// 1. "HARNESS_STARTED" might be used to indicate the worker harness has
+ /// started.
+ /// 2. "GCS_DOWNLOAD_ERROR" might be used to indicate an error downloading
+ /// a GCS file as part of the boot process of one of the worker containers.
+ ///
+ /// This is a string and not an enum to make it easy to add new codes without
+ /// waiting for an API change.
core.String code;
- /**
- * Parameters contains specific information about the code.
- *
- * This is a struct to allow parameters of different types.
- *
- * Examples:
- * 1. For a "HARNESS_STARTED" message parameters might provide the name
- * of the worker and additional data like timing information.
- * 2. For a "GCS_DOWNLOAD_ERROR" parameters might contain fields listing
- * the GCS objects being downloaded and fields containing errors.
- *
- * In general complex data structures should be avoided. If a worker
- * needs to send a specific and complicated data structure then please
- * consider defining a new proto and adding it to the data oneof in
- * WorkerMessageResponse.
- *
- * Conventions:
- * Parameters should only be used for information that isn't typically passed
- * as a label.
- * hostname and other worker identifiers should almost always be passed
- * as labels since they will be included on most messages.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Parameters contains specific information about the code.
+ ///
+ /// This is a struct to allow parameters of different types.
+ ///
+ /// Examples:
+ /// 1. For a "HARNESS_STARTED" message parameters might provide the name
+ /// of the worker and additional data like timing information.
+ /// 2. For a "GCS_DOWNLOAD_ERROR" parameters might contain fields listing
+ /// the GCS objects being downloaded and fields containing errors.
+ ///
+ /// In general complex data structures should be avoided. If a worker
+ /// needs to send a specific and complicated data structure then please
+ /// consider defining a new proto and adding it to the data oneof in
+ /// WorkerMessageResponse.
+ ///
+ /// Conventions:
+ /// Parameters should only be used for information that isn't typically
+ /// passed
+ /// as a label.
+ /// hostname and other worker identifiers should almost always be passed
+ /// as labels since they will be included on most messages.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> parameters;
WorkerMessageCode();
@@ -8596,7 +8962,8 @@ class WorkerMessageCode {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (code != null) {
_json["code"] = code;
}
@@ -8607,189 +8974,195 @@ class WorkerMessageCode {
}
}
-/**
- * A worker_message response allows the server to pass information to the
- * sender.
- */
+/// A worker_message response allows the server to pass information to the
+/// sender.
class WorkerMessageResponse {
- /** The service's response to a worker's health report. */
+ /// The service's response to a worker's health report.
WorkerHealthReportResponse workerHealthReportResponse;
- /** Service's response to reporting worker metrics (currently empty). */
+
+ /// Service's response to reporting worker metrics (currently empty).
ResourceUtilizationReportResponse workerMetricsResponse;
+ /// Service's response to shutdown notice (currently empty).
+ WorkerShutdownNoticeResponse workerShutdownNoticeResponse;
+
WorkerMessageResponse();
WorkerMessageResponse.fromJson(core.Map _json) {
if (_json.containsKey("workerHealthReportResponse")) {
- workerHealthReportResponse = new WorkerHealthReportResponse.fromJson(_json["workerHealthReportResponse"]);
+ workerHealthReportResponse = new WorkerHealthReportResponse.fromJson(
+ _json["workerHealthReportResponse"]);
}
if (_json.containsKey("workerMetricsResponse")) {
- workerMetricsResponse = new ResourceUtilizationReportResponse.fromJson(_json["workerMetricsResponse"]);
+ workerMetricsResponse = new ResourceUtilizationReportResponse.fromJson(
+ _json["workerMetricsResponse"]);
+ }
+ if (_json.containsKey("workerShutdownNoticeResponse")) {
+ workerShutdownNoticeResponse = new WorkerShutdownNoticeResponse.fromJson(
+ _json["workerShutdownNoticeResponse"]);
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (workerHealthReportResponse != null) {
- _json["workerHealthReportResponse"] = (workerHealthReportResponse).toJson();
+ _json["workerHealthReportResponse"] =
+ (workerHealthReportResponse).toJson();
}
if (workerMetricsResponse != null) {
_json["workerMetricsResponse"] = (workerMetricsResponse).toJson();
}
+ if (workerShutdownNoticeResponse != null) {
+ _json["workerShutdownNoticeResponse"] =
+ (workerShutdownNoticeResponse).toJson();
+ }
return _json;
}
}
-/**
- * Describes one particular pool of Cloud Dataflow workers to be
- * instantiated by the Cloud Dataflow service in order to perform the
- * computations required by a job. Note that a workflow job may use
- * multiple pools, in order to match the various computational
- * requirements of the various stages of the job.
- */
+/// Describes one particular pool of Cloud Dataflow workers to be
+/// instantiated by the Cloud Dataflow service in order to perform the
+/// computations required by a job. Note that a workflow job may use
+/// multiple pools, in order to match the various computational
+/// requirements of the various stages of the job.
class WorkerPool {
- /** Settings for autoscaling of this WorkerPool. */
+ /// Settings for autoscaling of this WorkerPool.
AutoscalingSettings autoscalingSettings;
- /** Data disks that are used by a VM in this workflow. */
+
+ /// Data disks that are used by a VM in this workflow.
core.List<Disk> dataDisks;
- /**
- * The default package set to install. This allows the service to
- * select a default set of packages which are useful to worker
- * harnesses written in a particular language.
- * Possible string values are:
- * - "DEFAULT_PACKAGE_SET_UNKNOWN" : The default set of packages to stage is
- * unknown, or unspecified.
- * - "DEFAULT_PACKAGE_SET_NONE" : Indicates that no packages should be staged
- * at the worker unless
- * explicitly specified by the job.
- * - "DEFAULT_PACKAGE_SET_JAVA" : Stage packages typically useful to workers
- * written in Java.
- * - "DEFAULT_PACKAGE_SET_PYTHON" : Stage pacakges typically useful to workers
- * written in Python.
- */
+
+ /// The default package set to install. This allows the service to
+ /// select a default set of packages which are useful to worker
+ /// harnesses written in a particular language.
+ /// Possible string values are:
+ /// - "DEFAULT_PACKAGE_SET_UNKNOWN" : The default set of packages to stage is
+ /// unknown, or unspecified.
+ /// - "DEFAULT_PACKAGE_SET_NONE" : Indicates that no packages should be
+ /// staged at the worker unless
+ /// explicitly specified by the job.
+ /// - "DEFAULT_PACKAGE_SET_JAVA" : Stage packages typically useful to workers
+ /// written in Java.
+ /// - "DEFAULT_PACKAGE_SET_PYTHON" : Stage pacakges typically useful to
+ /// workers written in Python.
core.String defaultPackageSet;
- /**
- * Size of root disk for VMs, in GB. If zero or unspecified, the service will
- * attempt to choose a reasonable default.
- */
+
+ /// Size of root disk for VMs, in GB. If zero or unspecified, the service
+ /// will
+ /// attempt to choose a reasonable default.
core.int diskSizeGb;
- /** Fully qualified source image for disks. */
+
+ /// Fully qualified source image for disks.
core.String diskSourceImage;
- /**
- * Type of root disk for VMs. If empty or unspecified, the service will
- * attempt to choose a reasonable default.
- */
+
+ /// Type of root disk for VMs. If empty or unspecified, the service will
+ /// attempt to choose a reasonable default.
core.String diskType;
- /**
- * Configuration for VM IPs.
- * Possible string values are:
- * - "WORKER_IP_UNSPECIFIED" : The configuration is unknown, or unspecified.
- * - "WORKER_IP_PUBLIC" : Workers should have public IP addresses.
- * - "WORKER_IP_PRIVATE" : Workers should have private IP addresses.
- */
+
+ /// Configuration for VM IPs.
+ /// Possible string values are:
+ /// - "WORKER_IP_UNSPECIFIED" : The configuration is unknown, or unspecified.
+ /// - "WORKER_IP_PUBLIC" : Workers should have public IP addresses.
+ /// - "WORKER_IP_PRIVATE" : Workers should have private IP addresses.
core.String ipConfiguration;
- /**
- * The kind of the worker pool; currently only `harness` and `shuffle`
- * are supported.
- */
+
+ /// The kind of the worker pool; currently only `harness` and `shuffle`
+ /// are supported.
core.String kind;
- /**
- * Machine type (e.g. "n1-standard-1"). If empty or unspecified, the
- * service will attempt to choose a reasonable default.
- */
+
+ /// Machine type (e.g. "n1-standard-1"). If empty or unspecified, the
+ /// service will attempt to choose a reasonable default.
core.String machineType;
- /** Metadata to set on the Google Compute Engine VMs. */
+
+ /// Metadata to set on the Google Compute Engine VMs.
core.Map<core.String, core.String> metadata;
- /**
- * Network to which VMs will be assigned. If empty or unspecified,
- * the service will use the network "default".
- */
+
+ /// Network to which VMs will be assigned. If empty or unspecified,
+ /// the service will use the network "default".
core.String network;
- /**
- * The number of threads per worker harness. If empty or unspecified, the
- * service will choose a number of threads (according to the number of cores
- * on the selected machine type for batch, or 1 by convention for streaming).
- */
+
+ /// The number of threads per worker harness. If empty or unspecified, the
+ /// service will choose a number of threads (according to the number of cores
+ /// on the selected machine type for batch, or 1 by convention for
+ /// streaming).
core.int numThreadsPerWorker;
- /**
- * Number of Google Compute Engine workers in this pool needed to
- * execute the job. If zero or unspecified, the service will
- * attempt to choose a reasonable default.
- */
+
+ /// Number of Google Compute Engine workers in this pool needed to
+ /// execute the job. If zero or unspecified, the service will
+ /// attempt to choose a reasonable default.
core.int numWorkers;
- /**
- * The action to take on host maintenance, as defined by the Google
- * Compute Engine API.
- */
+
+ /// The action to take on host maintenance, as defined by the Google
+ /// Compute Engine API.
core.String onHostMaintenance;
- /** Packages to be installed on workers. */
+
+ /// Packages to be installed on workers.
core.List<Package> packages;
- /**
- * Extra arguments for this worker pool.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Extra arguments for this worker pool.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> poolArgs;
- /**
- * Subnetwork to which VMs will be assigned, if desired. Expected to be of
- * the form "regions/REGION/subnetworks/SUBNETWORK".
- */
+
+ /// Subnetwork to which VMs will be assigned, if desired. Expected to be of
+ /// the form "regions/REGION/subnetworks/SUBNETWORK".
core.String subnetwork;
- /**
- * Settings passed through to Google Compute Engine workers when
- * using the standard Dataflow task runner. Users should ignore
- * this field.
- */
+
+ /// Settings passed through to Google Compute Engine workers when
+ /// using the standard Dataflow task runner. Users should ignore
+ /// this field.
TaskRunnerSettings taskrunnerSettings;
- /**
- * Sets the policy for determining when to turndown worker pool.
- * Allowed values are: `TEARDOWN_ALWAYS`, `TEARDOWN_ON_SUCCESS`, and
- * `TEARDOWN_NEVER`.
- * `TEARDOWN_ALWAYS` means workers are always torn down regardless of whether
- * the job succeeds. `TEARDOWN_ON_SUCCESS` means workers are torn down
- * if the job succeeds. `TEARDOWN_NEVER` means the workers are never torn
- * down.
- *
- * If the workers are not torn down by the service, they will
- * continue to run and use Google Compute Engine VM resources in the
- * user's project until they are explicitly terminated by the user.
- * Because of this, Google recommends using the `TEARDOWN_ALWAYS`
- * policy except for small, manually supervised test jobs.
- *
- * If unknown or unspecified, the service will attempt to choose a reasonable
- * default.
- * Possible string values are:
- * - "TEARDOWN_POLICY_UNKNOWN" : The teardown policy isn't specified, or is
- * unknown.
- * - "TEARDOWN_ALWAYS" : Always teardown the resource.
- * - "TEARDOWN_ON_SUCCESS" : Teardown the resource on success. This is useful
- * for debugging
- * failures.
- * - "TEARDOWN_NEVER" : Never teardown the resource. This is useful for
- * debugging and
- * development.
- */
+
+ /// Sets the policy for determining when to turndown worker pool.
+ /// Allowed values are: `TEARDOWN_ALWAYS`, `TEARDOWN_ON_SUCCESS`, and
+ /// `TEARDOWN_NEVER`.
+ /// `TEARDOWN_ALWAYS` means workers are always torn down regardless of
+ /// whether
+ /// the job succeeds. `TEARDOWN_ON_SUCCESS` means workers are torn down
+ /// if the job succeeds. `TEARDOWN_NEVER` means the workers are never torn
+ /// down.
+ ///
+ /// If the workers are not torn down by the service, they will
+ /// continue to run and use Google Compute Engine VM resources in the
+ /// user's project until they are explicitly terminated by the user.
+ /// Because of this, Google recommends using the `TEARDOWN_ALWAYS`
+ /// policy except for small, manually supervised test jobs.
+ ///
+ /// If unknown or unspecified, the service will attempt to choose a
+ /// reasonable
+ /// default.
+ /// Possible string values are:
+ /// - "TEARDOWN_POLICY_UNKNOWN" : The teardown policy isn't specified, or is
+ /// unknown.
+ /// - "TEARDOWN_ALWAYS" : Always teardown the resource.
+ /// - "TEARDOWN_ON_SUCCESS" : Teardown the resource on success. This is
+ /// useful for debugging
+ /// failures.
+ /// - "TEARDOWN_NEVER" : Never teardown the resource. This is useful for
+ /// debugging and
+ /// development.
core.String teardownPolicy;
- /**
- * Required. Docker container image that executes the Cloud Dataflow worker
- * harness, residing in Google Container Registry.
- */
+
+ /// Required. Docker container image that executes the Cloud Dataflow worker
+ /// harness, residing in Google Container Registry.
core.String workerHarnessContainerImage;
- /**
- * Zone to run the worker pools in. If empty or unspecified, the service
- * will attempt to choose a reasonable default.
- */
+
+ /// Zone to run the worker pools in. If empty or unspecified, the service
+ /// will attempt to choose a reasonable default.
core.String zone;
WorkerPool();
WorkerPool.fromJson(core.Map _json) {
if (_json.containsKey("autoscalingSettings")) {
- autoscalingSettings = new AutoscalingSettings.fromJson(_json["autoscalingSettings"]);
+ autoscalingSettings =
+ new AutoscalingSettings.fromJson(_json["autoscalingSettings"]);
}
if (_json.containsKey("dataDisks")) {
- dataDisks = _json["dataDisks"].map((value) => new Disk.fromJson(value)).toList();
+ dataDisks =
+ _json["dataDisks"].map((value) => new Disk.fromJson(value)).toList();
}
if (_json.containsKey("defaultPackageSet")) {
defaultPackageSet = _json["defaultPackageSet"];
@@ -8828,7 +9201,9 @@ class WorkerPool {
onHostMaintenance = _json["onHostMaintenance"];
}
if (_json.containsKey("packages")) {
- packages = _json["packages"].map((value) => new Package.fromJson(value)).toList();
+ packages = _json["packages"]
+ .map((value) => new Package.fromJson(value))
+ .toList();
}
if (_json.containsKey("poolArgs")) {
poolArgs = _json["poolArgs"];
@@ -8837,7 +9212,8 @@ class WorkerPool {
subnetwork = _json["subnetwork"];
}
if (_json.containsKey("taskrunnerSettings")) {
- taskrunnerSettings = new TaskRunnerSettings.fromJson(_json["taskrunnerSettings"]);
+ taskrunnerSettings =
+ new TaskRunnerSettings.fromJson(_json["taskrunnerSettings"]);
}
if (_json.containsKey("teardownPolicy")) {
teardownPolicy = _json["teardownPolicy"];
@@ -8851,7 +9227,8 @@ class WorkerPool {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (autoscalingSettings != null) {
_json["autoscalingSettings"] = (autoscalingSettings).toJson();
}
@@ -8919,45 +9296,42 @@ class WorkerPool {
}
}
-/** Provides data to pass through to the worker harness. */
+/// Provides data to pass through to the worker harness.
class WorkerSettings {
- /**
- * The base URL for accessing Google Cloud APIs.
- *
- * When workers access Google Cloud APIs, they logically do so via
- * relative URLs. If this field is specified, it supplies the base
- * URL to use for resolving these relative URLs. The normative
- * algorithm used is defined by RFC 1808, "Relative Uniform Resource
- * Locators".
- *
- * If not specified, the default value is "http://www.googleapis.com/"
- */
+ /// The base URL for accessing Google Cloud APIs.
+ ///
+ /// When workers access Google Cloud APIs, they logically do so via
+ /// relative URLs. If this field is specified, it supplies the base
+ /// URL to use for resolving these relative URLs. The normative
+ /// algorithm used is defined by RFC 1808, "Relative Uniform Resource
+ /// Locators".
+ ///
+ /// If not specified, the default value is "http://www.googleapis.com/"
core.String baseUrl;
- /** Whether to send work progress updates to the service. */
+
+ /// Whether to send work progress updates to the service.
core.bool reportingEnabled;
- /**
- * The Cloud Dataflow service path relative to the root URL, for example,
- * "dataflow/v1b3/projects".
- */
+
+ /// The Cloud Dataflow service path relative to the root URL, for example,
+ /// "dataflow/v1b3/projects".
core.String servicePath;
- /**
- * The Shuffle service path relative to the root URL, for example,
- * "shuffle/v1beta1".
- */
+
+ /// The Shuffle service path relative to the root URL, for example,
+ /// "shuffle/v1beta1".
core.String shuffleServicePath;
- /**
- * The prefix of the resources the system should use for temporary
- * storage.
- *
- * The supported resource type is:
- *
- * Google Cloud Storage:
- *
- * storage.googleapis.com/{bucket}/{object}
- * bucket.storage.googleapis.com/{object}
- */
+
+ /// The prefix of the resources the system should use for temporary
+ /// storage.
+ ///
+ /// The supported resource type is:
+ ///
+ /// Google Cloud Storage:
+ ///
+ /// storage.googleapis.com/{bucket}/{object}
+ /// bucket.storage.googleapis.com/{object}
core.String tempStoragePrefix;
- /** The ID of the worker running this pipeline. */
+
+ /// The ID of the worker running this pipeline.
core.String workerId;
WorkerSettings();
@@ -8984,7 +9358,8 @@ class WorkerSettings {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (baseUrl != null) {
_json["baseUrl"] = baseUrl;
}
@@ -9007,14 +9382,54 @@ class WorkerSettings {
}
}
-/**
- * An instruction that writes records.
- * Takes one input, produces no outputs.
- */
+/// Shutdown notification from workers. This is to be sent by the shutdown
+/// script of the worker VM so that the backend knows that the VM is being
+/// shut down.
+class WorkerShutdownNotice {
+ /// Optional reason to be attached for the shutdown notice.
+ /// For example: "PREEMPTION" would indicate the VM is being shut down
+ /// because
+ /// of preemption. Other possible reasons may be added in the future.
+ core.String reason;
+
+ WorkerShutdownNotice();
+
+ WorkerShutdownNotice.fromJson(core.Map _json) {
+ if (_json.containsKey("reason")) {
+ reason = _json["reason"];
+ }
+ }
+
+ core.Map<core.String, core.Object> toJson() {
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
+ if (reason != null) {
+ _json["reason"] = reason;
+ }
+ return _json;
+ }
+}
+
+/// Service-side response to WorkerMessage issuing shutdown notice.
+class WorkerShutdownNoticeResponse {
+ WorkerShutdownNoticeResponse();
+
+ WorkerShutdownNoticeResponse.fromJson(core.Map _json) {}
+
+ core.Map<core.String, core.Object> toJson() {
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
+ return _json;
+ }
+}
+
+/// An instruction that writes records.
+/// Takes one input, produces no outputs.
class WriteInstruction {
- /** The input. */
+ /// The input.
InstructionInput input;
- /** The sink to write to. */
+
+ /// The sink to write to.
Sink sink;
WriteInstruction();
@@ -9029,7 +9444,8 @@ class WriteInstruction {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (input != null) {
_json["input"] = (input).toJson();
}
« no previous file with comments | « generated/googleapis_beta/lib/clouduseraccounts/beta.dart ('k') | generated/googleapis_beta/lib/dlp/v2beta1.dart » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698