Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(964)

Unified Diff: generated/googleapis_beta/lib/genomics/v1beta.dart

Issue 559053002: Generate 0.1.0 version of googleapis/googleapis_beta (Closed) Base URL: git@github.com:dart-lang/googleapis.git@master
Patch Set: Created 6 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
Index: generated/googleapis_beta/lib/genomics/v1beta.dart
diff --git a/generated/googleapis_beta/lib/genomics/v1beta.dart b/generated/googleapis_beta/lib/genomics/v1beta.dart
new file mode 100644
index 0000000000000000000000000000000000000000..c74ffdaaecfac47fa8e9f51d5e0e276505550ecd
--- /dev/null
+++ b/generated/googleapis_beta/lib/genomics/v1beta.dart
@@ -0,0 +1,4284 @@
+library googleapis_beta.genomics.v1beta;
+
+import "dart:core" as core;
+import "dart:collection" as collection;
+import "dart:async" as async;
+import "dart:convert" as convert;
+
+import "package:crypto/crypto.dart" as crypto;
+import 'package:http/http.dart' as http;
+import '../src/common_internal.dart' as common_internal;
+import '../common/common.dart' as common;
+
+export '../common/common.dart' show ApiRequestError;
+export '../common/common.dart' show DetailedApiRequestError;
+
+/** Provides access to Genomics data. */
+class GenomicsApi {
+ /** View and manage your data in Google BigQuery */
+ static const BigqueryScope = "https://www.googleapis.com/auth/bigquery";
+
+ /** Manage your data in Google Cloud Storage */
+ static const DevstorageReadWriteScope = "https://www.googleapis.com/auth/devstorage.read_write";
+
+ /** View and manage Genomics data */
+ static const GenomicsScope = "https://www.googleapis.com/auth/genomics";
+
+ /** View Genomics data */
+ static const GenomicsReadonlyScope = "https://www.googleapis.com/auth/genomics.readonly";
+
+
+ final common_internal.ApiRequester _requester;
+
+ BeaconsResourceApi get beacons => new BeaconsResourceApi(_requester);
+ CallsetsResourceApi get callsets => new CallsetsResourceApi(_requester);
+ DatasetsResourceApi get datasets => new DatasetsResourceApi(_requester);
+ ExperimentalResourceApi get experimental => new ExperimentalResourceApi(_requester);
+ JobsResourceApi get jobs => new JobsResourceApi(_requester);
+ ReadsResourceApi get reads => new ReadsResourceApi(_requester);
+ ReadsetsResourceApi get readsets => new ReadsetsResourceApi(_requester);
+ VariantsResourceApi get variants => new VariantsResourceApi(_requester);
+
+ GenomicsApi(http.Client client) :
+ _requester = new common_internal.ApiRequester(client, "https://www.googleapis.com/", "/genomics/v1beta/");
+}
+
+
+/** Not documented yet. */
+class BeaconsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ BeaconsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * This is an experimental API that provides a Global Alliance for Genomics
+ * and Health Beacon. It may change at any time.
+ *
+ * Request parameters:
+ *
+ * [variantsetId] - The ID of the variantset to query over. It must be public.
+ * Private variantsets will return an unauthorized exception.
+ *
+ * [allele] - Required. The allele to look for ('A', 'C', 'G' or 'T').
+ *
+ * [contig] - Required. The contig to query over.
+ *
+ * [position] - Required. The 0-based position to query.
+ *
+ * Completes with a [Beacon].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Beacon> get(core.String variantsetId, {core.String allele, core.String contig, core.String position}) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (variantsetId == null) {
+ throw new core.ArgumentError("Parameter variantsetId is required.");
+ }
+ if (allele != null) {
+ _queryParams["allele"] = [allele];
+ }
+ if (contig != null) {
+ _queryParams["contig"] = [contig];
+ }
+ if (position != null) {
+ _queryParams["position"] = [position];
+ }
+
+
+ _url = 'beacons/' + common_internal.Escaper.ecapeVariable('$variantsetId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Beacon.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class CallsetsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ CallsetsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Creates a new callset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [Callset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Callset> create(Callset request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'callsets';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Callset.fromJson(data));
+ }
+
+ /**
+ * Deletes a callset.
+ *
+ * Request parameters:
+ *
+ * [callsetId] - The ID of the callset to be deleted.
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future delete(core.String callsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (callsetId == null) {
+ throw new core.ArgumentError("Parameter callsetId is required.");
+ }
+
+ _downloadOptions = null;
+
+ _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
+
+ var _response = _requester.request(_url,
+ "DELETE",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => null);
+ }
+
+ /**
+ * Gets a callset by ID.
+ *
+ * Request parameters:
+ *
+ * [callsetId] - The ID of the callset.
+ *
+ * Completes with a [Callset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Callset> get(core.String callsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (callsetId == null) {
+ throw new core.ArgumentError("Parameter callsetId is required.");
+ }
+
+
+ _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Callset.fromJson(data));
+ }
+
+ /**
+ * Updates a callset. This method supports patch semantics.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [callsetId] - The ID of the callset to be updated.
+ *
+ * Completes with a [Callset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Callset> patch(Callset request, core.String callsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (callsetId == null) {
+ throw new core.ArgumentError("Parameter callsetId is required.");
+ }
+
+
+ _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
+
+ var _response = _requester.request(_url,
+ "PATCH",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Callset.fromJson(data));
+ }
+
+ /**
+ * Gets a list of callsets matching the criteria.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [SearchCallsetsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<SearchCallsetsResponse> search(SearchCallsetsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'callsets/search';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new SearchCallsetsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a callset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [callsetId] - The ID of the callset to be updated.
+ *
+ * Completes with a [Callset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Callset> update(Callset request, core.String callsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (callsetId == null) {
+ throw new core.ArgumentError("Parameter callsetId is required.");
+ }
+
+
+ _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
+
+ var _response = _requester.request(_url,
+ "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Callset.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class DatasetsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ DatasetsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Creates a new dataset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [Dataset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Dataset> create(Dataset request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'datasets';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Dataset.fromJson(data));
+ }
+
+ /**
+ * Deletes a dataset.
+ *
+ * Request parameters:
+ *
+ * [datasetId] - The ID of the dataset to be deleted.
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future delete(core.String datasetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (datasetId == null) {
+ throw new core.ArgumentError("Parameter datasetId is required.");
+ }
+
+ _downloadOptions = null;
+
+ _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
+
+ var _response = _requester.request(_url,
+ "DELETE",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => null);
+ }
+
+ /**
+ * Gets a dataset by ID.
+ *
+ * Request parameters:
+ *
+ * [datasetId] - The ID of the dataset.
+ *
+ * Completes with a [Dataset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Dataset> get(core.String datasetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (datasetId == null) {
+ throw new core.ArgumentError("Parameter datasetId is required.");
+ }
+
+
+ _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Dataset.fromJson(data));
+ }
+
+ /**
+ * Lists all datasets.
+ *
+ * Request parameters:
+ *
+ * [maxResults] - The maximum number of results returned by this request.
+ *
+ * [pageToken] - The continuation token, which is used to page through large
+ * result sets. To get the next page of results, set this parameter to the
+ * value of "nextPageToken" from the previous response.
+ *
+ * [projectId] - Only return datasets which belong to this Google Developers
+ * Console project. Only accepts project numbers.
+ *
+ * Completes with a [ListDatasetsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ListDatasetsResponse> list({core.String maxResults, core.String pageToken, core.String projectId}) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (maxResults != null) {
+ _queryParams["maxResults"] = [maxResults];
+ }
+ if (pageToken != null) {
+ _queryParams["pageToken"] = [pageToken];
+ }
+ if (projectId != null) {
+ _queryParams["projectId"] = [projectId];
+ }
+
+
+ _url = 'datasets';
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ListDatasetsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a dataset. This method supports patch semantics.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [datasetId] - The ID of the dataset to be updated.
+ *
+ * Completes with a [Dataset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Dataset> patch(Dataset request, core.String datasetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (datasetId == null) {
+ throw new core.ArgumentError("Parameter datasetId is required.");
+ }
+
+
+ _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
+
+ var _response = _requester.request(_url,
+ "PATCH",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Dataset.fromJson(data));
+ }
+
+ /**
+ * Undeletes a dataset by restoring a dataset which was deleted via this API.
+ * This operation is only possible for a week after the deletion occurred.
+ *
+ * Request parameters:
+ *
+ * [datasetId] - The ID of the dataset to be undeleted.
+ *
+ * Completes with a [Dataset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Dataset> undelete(core.String datasetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (datasetId == null) {
+ throw new core.ArgumentError("Parameter datasetId is required.");
+ }
+
+
+ _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId') + '/undelete';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Dataset.fromJson(data));
+ }
+
+ /**
+ * Updates a dataset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [datasetId] - The ID of the dataset to be updated.
+ *
+ * Completes with a [Dataset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Dataset> update(Dataset request, core.String datasetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (datasetId == null) {
+ throw new core.ArgumentError("Parameter datasetId is required.");
+ }
+
+
+ _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
+
+ var _response = _requester.request(_url,
+ "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Dataset.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class ExperimentalResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ ExperimentalJobsResourceApi get jobs => new ExperimentalJobsResourceApi(_requester);
+
+ ExperimentalResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+}
+
+
+/** Not documented yet. */
+class ExperimentalJobsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ ExperimentalJobsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Creates and asynchronously runs an ad-hoc job. This is an experimental call
+ * and may be removed or changed at any time.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [ExperimentalCreateJobResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ExperimentalCreateJobResponse> create(ExperimentalCreateJobRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'experimental/jobs/create';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ExperimentalCreateJobResponse.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class JobsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ JobsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Cancels a job by ID. Note that it is possible for partial results to be
+ * generated and stored for cancelled jobs.
+ *
+ * Request parameters:
+ *
+ * [jobId] - Required. The ID of the job.
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future cancel(core.String jobId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (jobId == null) {
+ throw new core.ArgumentError("Parameter jobId is required.");
+ }
+
+ _downloadOptions = null;
+
+ _url = 'jobs/' + common_internal.Escaper.ecapeVariable('$jobId') + '/cancel';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => null);
+ }
+
+ /**
+ * Gets a job by ID.
+ *
+ * Request parameters:
+ *
+ * [jobId] - Required. The ID of the job.
+ *
+ * Completes with a [Job].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Job> get(core.String jobId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (jobId == null) {
+ throw new core.ArgumentError("Parameter jobId is required.");
+ }
+
+
+ _url = 'jobs/' + common_internal.Escaper.ecapeVariable('$jobId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Job.fromJson(data));
+ }
+
+ /**
+ * Gets a list of jobs matching the criteria.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [SearchJobsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<SearchJobsResponse> search(SearchJobsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'jobs/search';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new SearchJobsResponse.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class ReadsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ ReadsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Gets a list of reads for one or more readsets. Reads search operates over a
+ * genomic coordinate space of reference sequence & position defined over the
+ * reference sequences to which the requested readsets are aligned. If a
+ * target positional range is specified, search returns all reads whose
+ * alignment to the reference genome overlap the range. A query which
+ * specifies only readset IDs yields all reads in those readsets, including
+ * unmapped reads. All reads returned (including reads on subsequent pages)
+ * are ordered by genomic coordinate (reference sequence & position). Reads
+ * with equivalent genomic coordinates are returned in a deterministic order.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [SearchReadsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<SearchReadsResponse> search(SearchReadsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'reads/search';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new SearchReadsResponse.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class ReadsetsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ ReadsetsCoveragebucketsResourceApi get coveragebuckets => new ReadsetsCoveragebucketsResourceApi(_requester);
+
+ ReadsetsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Deletes a readset.
+ *
+ * Request parameters:
+ *
+ * [readsetId] - The ID of the readset to be deleted. The caller must have
+ * WRITE permissions to the dataset associated with this readset.
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future delete(core.String readsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (readsetId == null) {
+ throw new core.ArgumentError("Parameter readsetId is required.");
+ }
+
+ _downloadOptions = null;
+
+ _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
+
+ var _response = _requester.request(_url,
+ "DELETE",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => null);
+ }
+
+ /**
+ * Exports readsets to a BAM file in Google Cloud Storage. Note that currently
+ * there may be some differences between exported BAM files and the original
+ * BAM file at the time of import. In particular, comments in the input file
+ * header will not be preserved, and some custom tags will be converted to
+ * strings.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [ExportReadsetsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ExportReadsetsResponse> export(ExportReadsetsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'readsets/export';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ExportReadsetsResponse.fromJson(data));
+ }
+
+ /**
+ * Gets a readset by ID.
+ *
+ * Request parameters:
+ *
+ * [readsetId] - The ID of the readset.
+ *
+ * Completes with a [Readset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Readset> get(core.String readsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (readsetId == null) {
+ throw new core.ArgumentError("Parameter readsetId is required.");
+ }
+
+
+ _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Readset.fromJson(data));
+ }
+
+ /**
+ * Creates readsets by asynchronously importing the provided information. Note
+ * that currently comments in the input file header are not imported and some
+ * custom tags will be converted to strings, rather than preserving tag types.
+ * The caller must have WRITE permissions to the dataset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [ImportReadsetsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ImportReadsetsResponse> import(ImportReadsetsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'readsets/import';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ImportReadsetsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a readset. This method supports patch semantics.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [readsetId] - The ID of the readset to be updated. The caller must have
+ * WRITE permissions to the dataset associated with this readset.
+ *
+ * Completes with a [Readset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Readset> patch(Readset request, core.String readsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (readsetId == null) {
+ throw new core.ArgumentError("Parameter readsetId is required.");
+ }
+
+
+ _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
+
+ var _response = _requester.request(_url,
+ "PATCH",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Readset.fromJson(data));
+ }
+
+ /**
+ * Gets a list of readsets matching the criteria.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [SearchReadsetsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<SearchReadsetsResponse> search(SearchReadsetsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'readsets/search';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new SearchReadsetsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a readset.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [readsetId] - The ID of the readset to be updated. The caller must have
+ * WRITE permissions to the dataset associated with this readset.
+ *
+ * Completes with a [Readset].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Readset> update(Readset request, core.String readsetId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (readsetId == null) {
+ throw new core.ArgumentError("Parameter readsetId is required.");
+ }
+
+
+ _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
+
+ var _response = _requester.request(_url,
+ "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Readset.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class ReadsetsCoveragebucketsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ ReadsetsCoveragebucketsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Lists fixed width coverage buckets for a readset, each of which correspond
+ * to a range of a reference sequence. Each bucket summarizes coverage
+ * information across its corresponding genomic range. Coverage is defined as
+ * the number of reads which are aligned to a given base in the reference
+ * sequence. Coverage buckets are available at various bucket widths, enabling
+ * various coverage "zoom levels". The caller must have READ permissions for
+ * the target readset.
+ *
+ * Request parameters:
+ *
+ * [readsetId] - Required. The ID of the readset over which coverage is
+ * requested.
+ *
+ * [maxResults] - The maximum number of results to return in a single page. If
+ * unspecified, defaults to 1024. The maximum value is 2048.
+ *
+ * [pageToken] - The continuation token, which is used to page through large
+ * result sets. To get the next page of results, set this parameter to the
+ * value of "nextPageToken" from the previous response.
+ *
+ * [range_sequenceEnd] - The end position of the range on the reference,
+ * 1-based exclusive. If specified, sequenceName must also be specified.
+ *
+ * [range_sequenceName] - The reference sequence name, for example "chr1",
+ * "1", or "chrX".
+ *
+ * [range_sequenceStart] - The start position of the range on the reference,
+ * 1-based inclusive. If specified, sequenceName must also be specified.
+ *
+ * [targetBucketWidth] - The desired width of each reported coverage bucket in
+ * base pairs. This will be rounded down to the nearest precomputed bucket
+ * width; the value of which is returned as bucket_width in the response.
+ * Defaults to infinity (each bucket spans an entire reference sequence) or
+ * the length of the target range, if specified. The smallest precomputed
+ * bucket_width is currently 2048 base pairs; this is subject to change.
+ *
+ * Completes with a [ListCoverageBucketsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ListCoverageBucketsResponse> list(core.String readsetId, {core.String maxResults, core.String pageToken, core.String range_sequenceEnd, core.String range_sequenceName, core.String range_sequenceStart, core.String targetBucketWidth}) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (readsetId == null) {
+ throw new core.ArgumentError("Parameter readsetId is required.");
+ }
+ if (maxResults != null) {
+ _queryParams["maxResults"] = [maxResults];
+ }
+ if (pageToken != null) {
+ _queryParams["pageToken"] = [pageToken];
+ }
+ if (range_sequenceEnd != null) {
+ _queryParams["range.sequenceEnd"] = [range_sequenceEnd];
+ }
+ if (range_sequenceName != null) {
+ _queryParams["range.sequenceName"] = [range_sequenceName];
+ }
+ if (range_sequenceStart != null) {
+ _queryParams["range.sequenceStart"] = [range_sequenceStart];
+ }
+ if (targetBucketWidth != null) {
+ _queryParams["targetBucketWidth"] = [targetBucketWidth];
+ }
+
+
+ _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId') + '/coveragebuckets';
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ListCoverageBucketsResponse.fromJson(data));
+ }
+
+}
+
+
+/** Not documented yet. */
+class VariantsResourceApi {
+ final common_internal.ApiRequester _requester;
+
+ VariantsResourceApi(common_internal.ApiRequester client) :
+ _requester = client;
+
+ /**
+ * Creates a new variant.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [Variant].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Variant> create(Variant request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'variants';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Variant.fromJson(data));
+ }
+
+ /**
+ * Deletes a variant.
+ *
+ * Request parameters:
+ *
+ * [variantId] - The ID of the variant to be deleted.
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future delete(core.String variantId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (variantId == null) {
+ throw new core.ArgumentError("Parameter variantId is required.");
+ }
+
+ _downloadOptions = null;
+
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
+
+ var _response = _requester.request(_url,
+ "DELETE",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => null);
+ }
+
+ /**
+ * Exports variant data to an external destination.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [ExportVariantsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ExportVariantsResponse> export(ExportVariantsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'variants/export';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ExportVariantsResponse.fromJson(data));
+ }
+
+ /**
+ * Gets a variant by ID.
+ *
+ * Request parameters:
+ *
+ * [variantId] - The ID of the variant.
+ *
+ * Completes with a [Variant].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Variant> get(core.String variantId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (variantId == null) {
+ throw new core.ArgumentError("Parameter variantId is required.");
+ }
+
+
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Variant.fromJson(data));
+ }
+
+ /**
+ * Gets a summary of all the variant data in a variantset.
+ *
+ * Request parameters:
+ *
+ * [variantsetId] - Required. The ID of the variant set to get variant summary
+ * information for.
+ *
+ * Completes with a [GetVariantsSummaryResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<GetVariantsSummaryResponse> getSummary({core.String variantsetId}) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (variantsetId != null) {
+ _queryParams["variantsetId"] = [variantsetId];
+ }
+
+
+ _url = 'variants/summary';
+
+ var _response = _requester.request(_url,
+ "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new GetVariantsSummaryResponse.fromJson(data));
+ }
+
+ /**
+ * Creates variant data by asynchronously importing the provided information.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [ImportVariantsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<ImportVariantsResponse> import(ImportVariantsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'variants/import';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new ImportVariantsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a variant. This method supports patch semantics.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [variantId] - The ID of the variant to be updated..
+ *
+ * Completes with a [Variant].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Variant> patch(Variant request, core.String variantId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (variantId == null) {
+ throw new core.ArgumentError("Parameter variantId is required.");
+ }
+
+
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
+
+ var _response = _requester.request(_url,
+ "PATCH",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Variant.fromJson(data));
+ }
+
+ /**
+ * Gets a list of variants matching the criteria.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * Completes with a [SearchVariantsResponse].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<SearchVariantsResponse> search(SearchVariantsRequest request) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+
+
+ _url = 'variants/search';
+
+ var _response = _requester.request(_url,
+ "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new SearchVariantsResponse.fromJson(data));
+ }
+
+ /**
+ * Updates a variant.
+ *
+ * [request] - The metadata request object.
+ *
+ * Request parameters:
+ *
+ * [variantId] - The ID of the variant to be updated..
+ *
+ * Completes with a [Variant].
+ *
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an
+ * error.
+ *
+ * If the used [http.Client] completes with an error when making a REST call,
+ * this method will complete with the same error.
+ */
+ async.Future<Variant> update(Variant request, core.String variantId) {
+ var _url = null;
+ var _queryParams = new core.Map();
+ var _uploadMedia = null;
+ var _uploadOptions = null;
+ var _downloadOptions = common.DownloadOptions.Metadata;
+ var _body = null;
+
+ if (request != null) {
+ _body = convert.JSON.encode((request).toJson());
+ }
+ if (variantId == null) {
+ throw new core.ArgumentError("Parameter variantId is required.");
+ }
+
+
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
+
+ var _response = _requester.request(_url,
+ "PUT",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response.then((data) => new Variant.fromJson(data));
+ }
+
+}
+
+
+
+/**
+ * A beacon represents whether any variant call in a variantset has a specific
+ * allele at a particular position.
+ */
+class Beacon {
+ /** True if the allele exists on any variant call, false otherwise. */
+ core.bool exists;
+
+
+ Beacon();
+
+ Beacon.fromJson(core.Map _json) {
+ if (_json.containsKey("exists")) {
+ exists = _json["exists"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (exists != null) {
+ _json["exists"] = exists;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * A Call represents the determination of genotype with respect to a particular
+ * variant. It may include associated information such as quality and phasing.
+ * For example, a Call might assign a probability of 0.32 to the occurrence of a
+ * SNP named rs1234 in a callset with the name NA12345.
+ */
+class Call {
+ /** The ID of the callset this variant call belongs to. */
+ core.String callsetId;
+
+ /** The name of the callset this variant call belongs to. */
+ core.String callsetName;
+
+ /**
+ * The genotype of this variant call. Each value represents either the value
+ * of the referenceBases field or a 1-based index into alternateBases. If a
+ * variant had a referenceBases field of "T" and an alternateBases value of
+ * ["A", "C"], and the genotype was [2, 1], that would mean the call
+ * represented the heterozygous value "CA" for this variant. If the genotype
+ * was instead [0, 1], the represented value would be "TA". Ordering of the
+ * genotype values is important if the phaseset field is present. If a
+ * genotype is not called (that is, a "." is present in the GT string) -1 is
+ * returned.
+ */
+ core.List<core.int> genotype;
+
+ /**
+ * The genotype likelihoods for this variant call. Each array entry represents
+ * how likely a specific genotype is for this call. The value ordering is
+ * defined by the GL tag in the VCF spec.
+ */
+ core.List<core.double> genotypeLikelihood;
+
+ /** A map of additional variant call information. */
+ core.Map<core.String, core.List<core.String>> info;
+
+ /**
+ * If this field is present, this variant call's genotype ordering implies the
+ * phase of the bases and is consistent with any other variant calls on the
+ * same contig which have the same phaseset value. When importing data from
+ * VCF, if the genotype data was phased but no phase set was specified this
+ * field will be set to "*".
+ */
+ core.String phaseset;
+
+
+ Call();
+
+ Call.fromJson(core.Map _json) {
+ if (_json.containsKey("callsetId")) {
+ callsetId = _json["callsetId"];
+ }
+ if (_json.containsKey("callsetName")) {
+ callsetName = _json["callsetName"];
+ }
+ if (_json.containsKey("genotype")) {
+ genotype = _json["genotype"];
+ }
+ if (_json.containsKey("genotypeLikelihood")) {
+ genotypeLikelihood = _json["genotypeLikelihood"];
+ }
+ if (_json.containsKey("info")) {
+ info = _json["info"];
+ }
+ if (_json.containsKey("phaseset")) {
+ phaseset = _json["phaseset"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (callsetId != null) {
+ _json["callsetId"] = callsetId;
+ }
+ if (callsetName != null) {
+ _json["callsetName"] = callsetName;
+ }
+ if (genotype != null) {
+ _json["genotype"] = genotype;
+ }
+ if (genotypeLikelihood != null) {
+ _json["genotypeLikelihood"] = genotypeLikelihood;
+ }
+ if (info != null) {
+ _json["info"] = info;
+ }
+ if (phaseset != null) {
+ _json["phaseset"] = phaseset;
+ }
+ return _json;
+ }
+}
+
+
+/** A Callset is a collection of Variant Calls. It belongs to a Variantset. */
+class Callset {
+ /** The date this callset was created in milliseconds from the epoch. */
+ core.String created;
+
+ /** The Google generated ID of the callset, immutable. */
+ core.String id;
+
+ /** A map of additional callset information. */
+ core.Map<core.String, core.List<core.String>> info;
+
+ /** The callset name. */
+ core.String name;
+
+ /** The ID of the variantset this callset belongs to. */
+ core.String variantsetId;
+
+
+ Callset();
+
+ Callset.fromJson(core.Map _json) {
+ if (_json.containsKey("created")) {
+ created = _json["created"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("info")) {
+ info = _json["info"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("variantsetId")) {
+ variantsetId = _json["variantsetId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (created != null) {
+ _json["created"] = created;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (info != null) {
+ _json["info"] = info;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (variantsetId != null) {
+ _json["variantsetId"] = variantsetId;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * ContigBound records an upper bound for the starting coordinate of variants in
+ * a particular contig.
+ */
+class ContigBound {
+ /** The contig the bound is associate with. */
+ core.String contig;
+
+ /**
+ * An upper bound (inclusive) on the starting coordinate of any variant in the
+ * contig.
+ */
+ core.String upperBound;
+
+
+ ContigBound();
+
+ ContigBound.fromJson(core.Map _json) {
+ if (_json.containsKey("contig")) {
+ contig = _json["contig"];
+ }
+ if (_json.containsKey("upperBound")) {
+ upperBound = _json["upperBound"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (contig != null) {
+ _json["contig"] = contig;
+ }
+ if (upperBound != null) {
+ _json["upperBound"] = upperBound;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * A bucket over which read coverage has been precomputed. A bucket corresponds
+ * to a specific range of the reference sequence.
+ */
+class CoverageBucket {
+ /**
+ * The average number of reads which are aligned to each individual reference
+ * base in this bucket.
+ */
+ core.double meanCoverage;
+
+ /** The genomic coordinate range spanned by this bucket. */
+ GenomicRange range;
+
+
+ CoverageBucket();
+
+ CoverageBucket.fromJson(core.Map _json) {
+ if (_json.containsKey("meanCoverage")) {
+ meanCoverage = _json["meanCoverage"];
+ }
+ if (_json.containsKey("range")) {
+ range = new GenomicRange.fromJson(_json["range"]);
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (meanCoverage != null) {
+ _json["meanCoverage"] = meanCoverage;
+ }
+ if (range != null) {
+ _json["range"] = (range).toJson();
+ }
+ return _json;
+ }
+}
+
+
+/** A Dataset is a collection of genomic data. */
+class Dataset {
+ /** The Google generated ID of the dataset, immutable. */
+ core.String id;
+
+ /**
+ * Flag indicating whether or not a dataset is publicly viewable. If a dataset
+ * is not public, it inherits viewing permissions from its project.
+ */
+ core.bool isPublic;
+
+ /** The dataset name. */
+ core.String name;
+
+ /**
+ * The Google Developers Console project number that this dataset belongs to.
+ */
+ core.String projectId;
+
+
+ Dataset();
+
+ Dataset.fromJson(core.Map _json) {
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("isPublic")) {
+ isPublic = _json["isPublic"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (isPublic != null) {
+ _json["isPublic"] = isPublic;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ return _json;
+ }
+}
+
+
+/** The job creation request. */
+class ExperimentalCreateJobRequest {
+ /**
+ * Specifies whether or not to run the alignment pipeline. At least one of
+ * align or call_variants must be provided.
+ */
+ core.bool align;
+
+ /**
+ * Specifies whether or not to run the variant calling pipeline. If specified,
+ * alignment will be performed first and the aligned BAMs will passed as input
+ * to the variant caller. At least one of align or call_variants must be
+ * provided.
+ */
+ core.bool callVariants;
+
+ /**
+ * Specifies where to copy the results of certain pipelines. This shoud be in
+ * the form of "gs://bucket/path".
+ */
+ core.String gcsOutputPath;
+
+ /**
+ * A list of Google Cloud Storage URIs of paired end .fastq files to operate
+ * upon. If specified, this represents the second file of each paired .fastq
+ * file. The first file of each pair should be specified in "sourceUris".
+ */
+ core.List<core.String> pairedSourceUris;
+
+ /**
+ * Required. The Google Cloud Project ID with which to associate the request.
+ */
+ core.String projectId;
+
+ /**
+ * A list of Google Cloud Storage URIs of data files to operate upon. These
+ * can be .bam, interleaved .fastq, or paired .fastq. If specifying paired
+ * .fastq files, the first of each pair of files should be listed here, and
+ * the second of each pair should be listed in "pairedSourceUris".
+ */
+ core.List<core.String> sourceUris;
+
+
+ ExperimentalCreateJobRequest();
+
+ ExperimentalCreateJobRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("align")) {
+ align = _json["align"];
+ }
+ if (_json.containsKey("callVariants")) {
+ callVariants = _json["callVariants"];
+ }
+ if (_json.containsKey("gcsOutputPath")) {
+ gcsOutputPath = _json["gcsOutputPath"];
+ }
+ if (_json.containsKey("pairedSourceUris")) {
+ pairedSourceUris = _json["pairedSourceUris"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ if (_json.containsKey("sourceUris")) {
+ sourceUris = _json["sourceUris"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (align != null) {
+ _json["align"] = align;
+ }
+ if (callVariants != null) {
+ _json["callVariants"] = callVariants;
+ }
+ if (gcsOutputPath != null) {
+ _json["gcsOutputPath"] = gcsOutputPath;
+ }
+ if (pairedSourceUris != null) {
+ _json["pairedSourceUris"] = pairedSourceUris;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ if (sourceUris != null) {
+ _json["sourceUris"] = sourceUris;
+ }
+ return _json;
+ }
+}
+
+
+/** The job creation response. */
+class ExperimentalCreateJobResponse {
+ /** A job ID that can be used to get status information. */
+ core.String jobId;
+
+
+ ExperimentalCreateJobResponse();
+
+ ExperimentalCreateJobResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobId")) {
+ jobId = _json["jobId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobId != null) {
+ _json["jobId"] = jobId;
+ }
+ return _json;
+ }
+}
+
+
+/** The readset export request. */
+class ExportReadsetsRequest {
+ /**
+ * A Google Cloud Storage URI where the exported BAM file will be created. The
+ * currently authenticated user must have write access to the new file
+ * location. An error will be returned if the URI already contains data.
+ */
+ core.String exportUri;
+
+ /**
+ * The Google Developers Console project number that owns this export. This is
+ * the project that will be billed.
+ */
+ core.String projectId;
+
+ /** The IDs of the readsets to export. */
+ core.List<core.String> readsetIds;
+
+ /**
+ * The reference names to export. If this is not specified, all reference
+ * sequences are exported. Use '*' to export unmapped reads.
+ */
+ core.List<core.String> referenceNames;
+
+
+ ExportReadsetsRequest();
+
+ ExportReadsetsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("exportUri")) {
+ exportUri = _json["exportUri"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ if (_json.containsKey("readsetIds")) {
+ readsetIds = _json["readsetIds"];
+ }
+ if (_json.containsKey("referenceNames")) {
+ referenceNames = _json["referenceNames"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (exportUri != null) {
+ _json["exportUri"] = exportUri;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ if (readsetIds != null) {
+ _json["readsetIds"] = readsetIds;
+ }
+ if (referenceNames != null) {
+ _json["referenceNames"] = referenceNames;
+ }
+ return _json;
+ }
+}
+
+
+/** The readset export response. */
+class ExportReadsetsResponse {
+ /** A job ID that can be used to get status information. */
+ core.String jobId;
+
+
+ ExportReadsetsResponse();
+
+ ExportReadsetsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobId")) {
+ jobId = _json["jobId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobId != null) {
+ _json["jobId"] = jobId;
+ }
+ return _json;
+ }
+}
+
+
+/** The variant data export request. */
+class ExportVariantsRequest {
+ /**
+ * The BigQuery dataset to export data to. Note that this is distinct from the
+ * Genomics concept of "dataset". The caller must have WRITE access to this
+ * BigQuery dataset.
+ */
+ core.String bigqueryDataset;
+
+ /**
+ * The BigQuery table to export data to. The caller must have WRITE access to
+ * this BigQuery table.
+ */
+ core.String bigqueryTable;
+
+ /**
+ * If provided, only variant call information from the specified callsets will
+ * be exported. By default all variant calls are exported.
+ */
+ core.List<core.String> callsetIds;
+
+ /**
+ * The format for the exported data.
+ * Possible string values are:
+ * - "bigquery"
+ */
+ core.String format;
+
+ /**
+ * The Google Cloud project number that owns this export. This is the project
+ * that will be billed.
+ */
+ core.String projectId;
+
+ /**
+ * Required. The ID of the variantset that contains variant data which should
+ * be exported. The caller must have READ access to this variantset.
+ */
+ core.String variantsetId;
+
+
+ ExportVariantsRequest();
+
+ ExportVariantsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("bigqueryDataset")) {
+ bigqueryDataset = _json["bigqueryDataset"];
+ }
+ if (_json.containsKey("bigqueryTable")) {
+ bigqueryTable = _json["bigqueryTable"];
+ }
+ if (_json.containsKey("callsetIds")) {
+ callsetIds = _json["callsetIds"];
+ }
+ if (_json.containsKey("format")) {
+ format = _json["format"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ if (_json.containsKey("variantsetId")) {
+ variantsetId = _json["variantsetId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (bigqueryDataset != null) {
+ _json["bigqueryDataset"] = bigqueryDataset;
+ }
+ if (bigqueryTable != null) {
+ _json["bigqueryTable"] = bigqueryTable;
+ }
+ if (callsetIds != null) {
+ _json["callsetIds"] = callsetIds;
+ }
+ if (format != null) {
+ _json["format"] = format;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ if (variantsetId != null) {
+ _json["variantsetId"] = variantsetId;
+ }
+ return _json;
+ }
+}
+
+
+/** The variant data export response. */
+class ExportVariantsResponse {
+ /** A job ID that can be used to get status information. */
+ core.String jobId;
+
+
+ ExportVariantsResponse();
+
+ ExportVariantsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobId")) {
+ jobId = _json["jobId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobId != null) {
+ _json["jobId"] = jobId;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * An inclusive, exclusive genomic coordinate range over a reference sequence.
+ */
+class GenomicRange {
+ /**
+ * The end position of the range on the reference, 1-based exclusive. If
+ * specified, sequenceName must also be specified.
+ */
+ core.String sequenceEnd;
+
+ /** The reference sequence name, for example "chr1", "1", or "chrX". */
+ core.String sequenceName;
+
+ /**
+ * The start position of the range on the reference, 1-based inclusive. If
+ * specified, sequenceName must also be specified.
+ */
+ core.String sequenceStart;
+
+
+ GenomicRange();
+
+ GenomicRange.fromJson(core.Map _json) {
+ if (_json.containsKey("sequenceEnd")) {
+ sequenceEnd = _json["sequenceEnd"];
+ }
+ if (_json.containsKey("sequenceName")) {
+ sequenceName = _json["sequenceName"];
+ }
+ if (_json.containsKey("sequenceStart")) {
+ sequenceStart = _json["sequenceStart"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (sequenceEnd != null) {
+ _json["sequenceEnd"] = sequenceEnd;
+ }
+ if (sequenceName != null) {
+ _json["sequenceName"] = sequenceName;
+ }
+ if (sequenceStart != null) {
+ _json["sequenceStart"] = sequenceStart;
+ }
+ return _json;
+ }
+}
+
+
+/** The variants summary response. */
+class GetVariantsSummaryResponse {
+ /**
+ * A list of all contigs used by the variants in a dataset with associated
+ * coordinate upper bounds for each one.
+ */
+ core.List<ContigBound> contigBounds;
+
+ /** The metadata associated with this dataset. */
+ core.List<Metadata> metadata;
+
+
+ GetVariantsSummaryResponse();
+
+ GetVariantsSummaryResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("contigBounds")) {
+ contigBounds = _json["contigBounds"].map((value) => new ContigBound.fromJson(value)).toList();
+ }
+ if (_json.containsKey("metadata")) {
+ metadata = _json["metadata"].map((value) => new Metadata.fromJson(value)).toList();
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (contigBounds != null) {
+ _json["contigBounds"] = contigBounds.map((value) => (value).toJson()).toList();
+ }
+ if (metadata != null) {
+ _json["metadata"] = metadata.map((value) => (value).toJson()).toList();
+ }
+ return _json;
+ }
+}
+
+
+/** Not documented yet. */
+class Header {
+ /** (SO) Sorting order of alignments. */
+ core.String sortingOrder;
+
+ /** (VN) BAM format version. */
+ core.String version;
+
+
+ Header();
+
+ Header.fromJson(core.Map _json) {
+ if (_json.containsKey("sortingOrder")) {
+ sortingOrder = _json["sortingOrder"];
+ }
+ if (_json.containsKey("version")) {
+ version = _json["version"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (sortingOrder != null) {
+ _json["sortingOrder"] = sortingOrder;
+ }
+ if (version != null) {
+ _json["version"] = version;
+ }
+ return _json;
+ }
+}
+
+
+/** The header section of the BAM/SAM file. */
+class HeaderSection {
+ /** (@CO) One-line text comments. */
+ core.List<core.String> comments;
+
+ /**
+ * [Deprecated] This field is deprecated and will no longer be populated.
+ * Please use filename instead.
+ */
+ core.String fileUri;
+
+ /** The name of the file from which this data was imported. */
+ core.String filename;
+
+ /** (@HD) The header line. */
+ core.List<Header> headers;
+
+ /** (@PG) Programs. */
+ core.List<Program> programs;
+
+ /** (@RG) Read group. */
+ core.List<ReadGroup> readGroups;
+
+ /** (@SQ) Reference sequence dictionary. */
+ core.List<ReferenceSequence> refSequences;
+
+
+ HeaderSection();
+
+ HeaderSection.fromJson(core.Map _json) {
+ if (_json.containsKey("comments")) {
+ comments = _json["comments"];
+ }
+ if (_json.containsKey("fileUri")) {
+ fileUri = _json["fileUri"];
+ }
+ if (_json.containsKey("filename")) {
+ filename = _json["filename"];
+ }
+ if (_json.containsKey("headers")) {
+ headers = _json["headers"].map((value) => new Header.fromJson(value)).toList();
+ }
+ if (_json.containsKey("programs")) {
+ programs = _json["programs"].map((value) => new Program.fromJson(value)).toList();
+ }
+ if (_json.containsKey("readGroups")) {
+ readGroups = _json["readGroups"].map((value) => new ReadGroup.fromJson(value)).toList();
+ }
+ if (_json.containsKey("refSequences")) {
+ refSequences = _json["refSequences"].map((value) => new ReferenceSequence.fromJson(value)).toList();
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (comments != null) {
+ _json["comments"] = comments;
+ }
+ if (fileUri != null) {
+ _json["fileUri"] = fileUri;
+ }
+ if (filename != null) {
+ _json["filename"] = filename;
+ }
+ if (headers != null) {
+ _json["headers"] = headers.map((value) => (value).toJson()).toList();
+ }
+ if (programs != null) {
+ _json["programs"] = programs.map((value) => (value).toJson()).toList();
+ }
+ if (readGroups != null) {
+ _json["readGroups"] = readGroups.map((value) => (value).toJson()).toList();
+ }
+ if (refSequences != null) {
+ _json["refSequences"] = refSequences.map((value) => (value).toJson()).toList();
+ }
+ return _json;
+ }
+}
+
+
+/** The readset import request. */
+class ImportReadsetsRequest {
+ /**
+ * Required. The ID of the dataset these readsets will belong to. The caller
+ * must have WRITE permissions to this dataset.
+ */
+ core.String datasetId;
+
+ /** A list of URIs pointing at BAM files in Google Cloud Storage. */
+ core.List<core.String> sourceUris;
+
+
+ ImportReadsetsRequest();
+
+ ImportReadsetsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("datasetId")) {
+ datasetId = _json["datasetId"];
+ }
+ if (_json.containsKey("sourceUris")) {
+ sourceUris = _json["sourceUris"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (datasetId != null) {
+ _json["datasetId"] = datasetId;
+ }
+ if (sourceUris != null) {
+ _json["sourceUris"] = sourceUris;
+ }
+ return _json;
+ }
+}
+
+
+/** The readset import response. */
+class ImportReadsetsResponse {
+ /** A job ID that can be used to get status information. */
+ core.String jobId;
+
+
+ ImportReadsetsResponse();
+
+ ImportReadsetsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobId")) {
+ jobId = _json["jobId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobId != null) {
+ _json["jobId"] = jobId;
+ }
+ return _json;
+ }
+}
+
+
+/** The variant data import request. */
+class ImportVariantsRequest {
+ /**
+ * The format of the variant data being imported.
+ * Possible string values are:
+ * - "completeGenomics"
+ * - "vcf"
+ */
+ core.String format;
+
+ /**
+ * A list of URIs pointing at VCF files in Google Cloud Storage. See the VCF
+ * Specification for more details on the input format.
+ */
+ core.List<core.String> sourceUris;
+
+ /** Required. The variantset to which variant data should be imported. */
+ core.String variantsetId;
+
+
+ ImportVariantsRequest();
+
+ ImportVariantsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("format")) {
+ format = _json["format"];
+ }
+ if (_json.containsKey("sourceUris")) {
+ sourceUris = _json["sourceUris"];
+ }
+ if (_json.containsKey("variantsetId")) {
+ variantsetId = _json["variantsetId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (format != null) {
+ _json["format"] = format;
+ }
+ if (sourceUris != null) {
+ _json["sourceUris"] = sourceUris;
+ }
+ if (variantsetId != null) {
+ _json["variantsetId"] = variantsetId;
+ }
+ return _json;
+ }
+}
+
+
+/** The variant data import response. */
+class ImportVariantsResponse {
+ /** A job ID that can be used to get status information. */
+ core.String jobId;
+
+
+ ImportVariantsResponse();
+
+ ImportVariantsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobId")) {
+ jobId = _json["jobId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobId != null) {
+ _json["jobId"] = jobId;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * A Job represents an ongoing process that can be monitored for status
+ * information.
+ */
+class Job {
+ /** The date this job was created, in milliseconds from the epoch. */
+ core.String created;
+
+ /** A more detailed description of this job's current status. */
+ core.String description;
+
+ /** Any errors that occurred during processing. */
+ core.List<core.String> errors;
+
+ /** The job ID. */
+ core.String id;
+
+ /**
+ * If this Job represents an import, this field will contain the IDs of the
+ * objects that were successfully imported.
+ */
+ core.List<core.String> importedIds;
+
+ /**
+ * The Google Developers Console project number to which this job belongs.
+ */
+ core.String projectId;
+
+ /** A summarized representation of the original service request. */
+ JobRequest request;
+
+ /**
+ * The status of this job.
+ * Possible string values are:
+ * - "canceled"
+ * - "failure"
+ * - "new"
+ * - "pending"
+ * - "running"
+ * - "success"
+ * - "unknownStatus"
+ */
+ core.String status;
+
+ /** Any warnings that occurred during processing. */
+ core.List<core.String> warnings;
+
+
+ Job();
+
+ Job.fromJson(core.Map _json) {
+ if (_json.containsKey("created")) {
+ created = _json["created"];
+ }
+ if (_json.containsKey("description")) {
+ description = _json["description"];
+ }
+ if (_json.containsKey("errors")) {
+ errors = _json["errors"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("importedIds")) {
+ importedIds = _json["importedIds"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ if (_json.containsKey("request")) {
+ request = new JobRequest.fromJson(_json["request"]);
+ }
+ if (_json.containsKey("status")) {
+ status = _json["status"];
+ }
+ if (_json.containsKey("warnings")) {
+ warnings = _json["warnings"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (created != null) {
+ _json["created"] = created;
+ }
+ if (description != null) {
+ _json["description"] = description;
+ }
+ if (errors != null) {
+ _json["errors"] = errors;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (importedIds != null) {
+ _json["importedIds"] = importedIds;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ if (request != null) {
+ _json["request"] = (request).toJson();
+ }
+ if (status != null) {
+ _json["status"] = status;
+ }
+ if (warnings != null) {
+ _json["warnings"] = warnings;
+ }
+ return _json;
+ }
+}
+
+
+/** A summary representation of the service request that spawned the job. */
+class JobRequest {
+ /**
+ * The data destination of the request, for example, a Google BigQuery Table
+ * or Dataset ID.
+ */
+ core.List<core.String> destination;
+
+ /**
+ * The data source of the request, for example, a Google Cloud Storage object
+ * path or Readset ID.
+ */
+ core.List<core.String> source;
+
+ /**
+ * The original request type.
+ * Possible string values are:
+ * - "experimentalCreateJob"
+ * - "exportReadsets"
+ * - "exportVariants"
+ * - "importReadsets"
+ * - "importVariants"
+ * - "unknownType"
+ */
+ core.String type;
+
+
+ JobRequest();
+
+ JobRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("destination")) {
+ destination = _json["destination"];
+ }
+ if (_json.containsKey("source")) {
+ source = _json["source"];
+ }
+ if (_json.containsKey("type")) {
+ type = _json["type"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (destination != null) {
+ _json["destination"] = destination;
+ }
+ if (source != null) {
+ _json["source"] = source;
+ }
+ if (type != null) {
+ _json["type"] = type;
+ }
+ return _json;
+ }
+}
+
+
+/** Not documented yet. */
+class ListCoverageBucketsResponse {
+ /**
+ * The length of each coverage bucket in base pairs. Note that buckets at the
+ * end of a reference sequence may be shorter. This value is omitted if the
+ * bucket width is infinity (the default behaviour, with no range or
+ * target_bucket_width).
+ */
+ core.String bucketWidth;
+
+ /**
+ * The coverage buckets. The list of buckets is sparse; a bucket with 0
+ * overlapping reads is not returned. A bucket never crosses more than one
+ * reference sequence. Each bucket has width bucket_width, unless its end is
+ * is the end of the reference sequence.
+ */
+ core.List<CoverageBucket> coverageBuckets;
+
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+
+ ListCoverageBucketsResponse();
+
+ ListCoverageBucketsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("bucketWidth")) {
+ bucketWidth = _json["bucketWidth"];
+ }
+ if (_json.containsKey("coverageBuckets")) {
+ coverageBuckets = _json["coverageBuckets"].map((value) => new CoverageBucket.fromJson(value)).toList();
+ }
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (bucketWidth != null) {
+ _json["bucketWidth"] = bucketWidth;
+ }
+ if (coverageBuckets != null) {
+ _json["coverageBuckets"] = coverageBuckets.map((value) => (value).toJson()).toList();
+ }
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ return _json;
+ }
+}
+
+
+/** The dataset list response. */
+class ListDatasetsResponse {
+ /** The list of matching Datasets. */
+ core.List<Dataset> datasets;
+
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+
+ ListDatasetsResponse();
+
+ ListDatasetsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("datasets")) {
+ datasets = _json["datasets"].map((value) => new Dataset.fromJson(value)).toList();
+ }
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (datasets != null) {
+ _json["datasets"] = datasets.map((value) => (value).toJson()).toList();
+ }
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * Metadata describes a single piece of variant call metadata. These data
+ * include a top level key and either a single value string (value) or a list of
+ * key-value pairs (info.) Value and info are mutually exclusive.
+ */
+class Metadata {
+ /** A textual description of this metadata. */
+ core.String description;
+
+ /**
+ * User-provided ID field, not enforced by this API. Two or more pieces of
+ * structured metadata with identical id and key fields are considered
+ * equivalent.
+ */
+ core.String id;
+
+ /** Remaining structured metadata key-value pairs. */
+ core.Map<core.String, core.List<core.String>> info;
+
+ /** The top-level key. */
+ core.String key;
+
+ /**
+ * The number of values that can be included in a field described by this
+ * metadata.
+ */
+ core.String number;
+
+ /**
+ * The type of data. Possible types include: Integer, Float, Flag, Character,
+ * and String.
+ * Possible string values are:
+ * - "character"
+ * - "flag"
+ * - "float"
+ * - "integer"
+ * - "string"
+ * - "unknownType"
+ */
+ core.String type;
+
+ /** The value field for simple metadata */
+ core.String value;
+
+
+ Metadata();
+
+ Metadata.fromJson(core.Map _json) {
+ if (_json.containsKey("description")) {
+ description = _json["description"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("info")) {
+ info = _json["info"];
+ }
+ if (_json.containsKey("key")) {
+ key = _json["key"];
+ }
+ if (_json.containsKey("number")) {
+ number = _json["number"];
+ }
+ if (_json.containsKey("type")) {
+ type = _json["type"];
+ }
+ if (_json.containsKey("value")) {
+ value = _json["value"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (description != null) {
+ _json["description"] = description;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (info != null) {
+ _json["info"] = info;
+ }
+ if (key != null) {
+ _json["key"] = key;
+ }
+ if (number != null) {
+ _json["number"] = number;
+ }
+ if (type != null) {
+ _json["type"] = type;
+ }
+ if (value != null) {
+ _json["value"] = value;
+ }
+ return _json;
+ }
+}
+
+
+/** Not documented yet. */
+class Program {
+ /** (CL) Command line. */
+ core.String commandLine;
+
+ /** (ID) Program record identifier. */
+ core.String id;
+
+ /** (PN) Program name. */
+ core.String name;
+
+ /** (PP) Previous program ID. */
+ core.String prevProgramId;
+
+ /** (VN) Program version. */
+ core.String version;
+
+
+ Program();
+
+ Program.fromJson(core.Map _json) {
+ if (_json.containsKey("commandLine")) {
+ commandLine = _json["commandLine"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("prevProgramId")) {
+ prevProgramId = _json["prevProgramId"];
+ }
+ if (_json.containsKey("version")) {
+ version = _json["version"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (commandLine != null) {
+ _json["commandLine"] = commandLine;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (prevProgramId != null) {
+ _json["prevProgramId"] = prevProgramId;
+ }
+ if (version != null) {
+ _json["version"] = version;
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * A Read is a group of bases that may or may not have been aligned to a
+ * reference. It contains quality information and other metadata.
+ */
+class Read {
+ /**
+ * The originalBases after the cigar field has been applied. Deletions are
+ * represented with '-' and insertions are omitted.
+ */
+ core.String alignedBases;
+
+ /**
+ * Represents the quality of each base in this read. Each character represents
+ * one base. To get the quality, take the ASCII value of the character and
+ * subtract 33. (QUAL)
+ */
+ core.String baseQuality;
+
+ /**
+ * A condensed representation of how this read matches up to the reference.
+ * (CIGAR)
+ */
+ core.String cigar;
+
+ /**
+ * Each bit of this number has a different meaning if enabled. See the full
+ * BAM spec for more details. (FLAG)
+ */
+ core.int flags;
+
+ /** The Google generated ID of the read, immutable. */
+ core.String id;
+
+ /**
+ * A score up to 255 that represents how likely this read's aligned position
+ * is to be correct. A higher value is better. (MAPQ)
+ */
+ core.int mappingQuality;
+
+ /** The 1-based start position of the paired read. (PNEXT) */
+ core.int matePosition;
+
+ /**
+ * The name of the sequence that the paired read is aligned to. This is
+ * usually the same as referenceSequenceName. (RNEXT)
+ */
+ core.String mateReferenceSequenceName;
+
+ /**
+ * The name of the read. When imported from a BAM file, this is the query
+ * template name. (QNAME)
+ */
+ core.String name;
+
+ /** The list of bases that this read represents (such as "CATCGA"). (SEQ) */
+ core.String originalBases;
+
+ /**
+ * The 1-based start position of the aligned read. If the first base starts at
+ * the very beginning of the reference sequence, then the position would be
+ * '1'. (POS)
+ */
+ core.int position;
+
+ /** The ID of the readset this read belongs to. */
+ core.String readsetId;
+
+ /**
+ * The name of the sequence that this read is aligned to. This would be, for
+ * example, 'X' for the X Chromosome or '20' for Chromosome 20. (RNAME)
+ */
+ core.String referenceSequenceName;
+
+ /** A map of additional read information. (TAG) */
+ core.Map<core.String, core.List<core.String>> tags;
+
+ /**
+ * Length of the original piece of DNA that produced both this read and the
+ * paired read. (TLEN)
+ */
+ core.int templateLength;
+
+
+ Read();
+
+ Read.fromJson(core.Map _json) {
+ if (_json.containsKey("alignedBases")) {
+ alignedBases = _json["alignedBases"];
+ }
+ if (_json.containsKey("baseQuality")) {
+ baseQuality = _json["baseQuality"];
+ }
+ if (_json.containsKey("cigar")) {
+ cigar = _json["cigar"];
+ }
+ if (_json.containsKey("flags")) {
+ flags = _json["flags"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("mappingQuality")) {
+ mappingQuality = _json["mappingQuality"];
+ }
+ if (_json.containsKey("matePosition")) {
+ matePosition = _json["matePosition"];
+ }
+ if (_json.containsKey("mateReferenceSequenceName")) {
+ mateReferenceSequenceName = _json["mateReferenceSequenceName"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("originalBases")) {
+ originalBases = _json["originalBases"];
+ }
+ if (_json.containsKey("position")) {
+ position = _json["position"];
+ }
+ if (_json.containsKey("readsetId")) {
+ readsetId = _json["readsetId"];
+ }
+ if (_json.containsKey("referenceSequenceName")) {
+ referenceSequenceName = _json["referenceSequenceName"];
+ }
+ if (_json.containsKey("tags")) {
+ tags = _json["tags"];
+ }
+ if (_json.containsKey("templateLength")) {
+ templateLength = _json["templateLength"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (alignedBases != null) {
+ _json["alignedBases"] = alignedBases;
+ }
+ if (baseQuality != null) {
+ _json["baseQuality"] = baseQuality;
+ }
+ if (cigar != null) {
+ _json["cigar"] = cigar;
+ }
+ if (flags != null) {
+ _json["flags"] = flags;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (mappingQuality != null) {
+ _json["mappingQuality"] = mappingQuality;
+ }
+ if (matePosition != null) {
+ _json["matePosition"] = matePosition;
+ }
+ if (mateReferenceSequenceName != null) {
+ _json["mateReferenceSequenceName"] = mateReferenceSequenceName;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (originalBases != null) {
+ _json["originalBases"] = originalBases;
+ }
+ if (position != null) {
+ _json["position"] = position;
+ }
+ if (readsetId != null) {
+ _json["readsetId"] = readsetId;
+ }
+ if (referenceSequenceName != null) {
+ _json["referenceSequenceName"] = referenceSequenceName;
+ }
+ if (tags != null) {
+ _json["tags"] = tags;
+ }
+ if (templateLength != null) {
+ _json["templateLength"] = templateLength;
+ }
+ return _json;
+ }
+}
+
+
+/** Not documented yet. */
+class ReadGroup {
+ /** (DT) Date the run was produced (ISO8601 date or date/time). */
+ core.String date;
+
+ /** (DS) Description. */
+ core.String description;
+
+ /**
+ * (FO) Flow order. The array of nucleotide bases that correspond to the
+ * nucleotides used for each flow of each read.
+ */
+ core.String flowOrder;
+
+ /** (ID) Read group identifier. */
+ core.String id;
+
+ /**
+ * (KS) The array of nucleotide bases that correspond to the key sequence of
+ * each read.
+ */
+ core.String keySequence;
+
+ /** (LS) Library. */
+ core.String library;
+
+ /** (PU) Platform unit. */
+ core.String platformUnit;
+
+ /** (PI) Predicted median insert size. */
+ core.int predictedInsertSize;
+
+ /** (PG) Programs used for processing the read group. */
+ core.String processingProgram;
+
+ /** (SM) Sample. */
+ core.String sample;
+
+ /** (CN) Name of sequencing center producing the read. */
+ core.String sequencingCenterName;
+
+ /** (PL) Platform/technology used to produce the reads. */
+ core.String sequencingTechnology;
+
+
+ ReadGroup();
+
+ ReadGroup.fromJson(core.Map _json) {
+ if (_json.containsKey("date")) {
+ date = _json["date"];
+ }
+ if (_json.containsKey("description")) {
+ description = _json["description"];
+ }
+ if (_json.containsKey("flowOrder")) {
+ flowOrder = _json["flowOrder"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("keySequence")) {
+ keySequence = _json["keySequence"];
+ }
+ if (_json.containsKey("library")) {
+ library = _json["library"];
+ }
+ if (_json.containsKey("platformUnit")) {
+ platformUnit = _json["platformUnit"];
+ }
+ if (_json.containsKey("predictedInsertSize")) {
+ predictedInsertSize = _json["predictedInsertSize"];
+ }
+ if (_json.containsKey("processingProgram")) {
+ processingProgram = _json["processingProgram"];
+ }
+ if (_json.containsKey("sample")) {
+ sample = _json["sample"];
+ }
+ if (_json.containsKey("sequencingCenterName")) {
+ sequencingCenterName = _json["sequencingCenterName"];
+ }
+ if (_json.containsKey("sequencingTechnology")) {
+ sequencingTechnology = _json["sequencingTechnology"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (date != null) {
+ _json["date"] = date;
+ }
+ if (description != null) {
+ _json["description"] = description;
+ }
+ if (flowOrder != null) {
+ _json["flowOrder"] = flowOrder;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (keySequence != null) {
+ _json["keySequence"] = keySequence;
+ }
+ if (library != null) {
+ _json["library"] = library;
+ }
+ if (platformUnit != null) {
+ _json["platformUnit"] = platformUnit;
+ }
+ if (predictedInsertSize != null) {
+ _json["predictedInsertSize"] = predictedInsertSize;
+ }
+ if (processingProgram != null) {
+ _json["processingProgram"] = processingProgram;
+ }
+ if (sample != null) {
+ _json["sample"] = sample;
+ }
+ if (sequencingCenterName != null) {
+ _json["sequencingCenterName"] = sequencingCenterName;
+ }
+ if (sequencingTechnology != null) {
+ _json["sequencingTechnology"] = sequencingTechnology;
+ }
+ return _json;
+ }
+}
+
+
+/** A Readset is a collection of Reads. */
+class Readset {
+ /** The ID of the dataset this readset belongs to. */
+ core.String datasetId;
+
+ /**
+ * File information from the original BAM import. See the BAM format
+ * specification for additional information on each field.
+ */
+ core.List<HeaderSection> fileData;
+
+ /** The Google generated ID of the readset, immutable. */
+ core.String id;
+
+ /** The readset name, typically this is the sample name. */
+ core.String name;
+
+
+ Readset();
+
+ Readset.fromJson(core.Map _json) {
+ if (_json.containsKey("datasetId")) {
+ datasetId = _json["datasetId"];
+ }
+ if (_json.containsKey("fileData")) {
+ fileData = _json["fileData"].map((value) => new HeaderSection.fromJson(value)).toList();
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (datasetId != null) {
+ _json["datasetId"] = datasetId;
+ }
+ if (fileData != null) {
+ _json["fileData"] = fileData.map((value) => (value).toJson()).toList();
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ return _json;
+ }
+}
+
+
+/** Not documented yet. */
+class ReferenceSequence {
+ /** (AS) Genome assembly identifier. */
+ core.String assemblyId;
+
+ /** (LN) Reference sequence length. */
+ core.int length;
+
+ /**
+ * (M5) MD5 checksum of the sequence in the uppercase, excluding spaces but
+ * including pads as *.
+ */
+ core.String md5Checksum;
+
+ /** (SN) Reference sequence name. */
+ core.String name;
+
+ /** (SP) Species. */
+ core.String species;
+
+ /** (UR) URI of the sequence. */
+ core.String uri;
+
+
+ ReferenceSequence();
+
+ ReferenceSequence.fromJson(core.Map _json) {
+ if (_json.containsKey("assemblyId")) {
+ assemblyId = _json["assemblyId"];
+ }
+ if (_json.containsKey("length")) {
+ length = _json["length"];
+ }
+ if (_json.containsKey("md5Checksum")) {
+ md5Checksum = _json["md5Checksum"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("species")) {
+ species = _json["species"];
+ }
+ if (_json.containsKey("uri")) {
+ uri = _json["uri"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (assemblyId != null) {
+ _json["assemblyId"] = assemblyId;
+ }
+ if (length != null) {
+ _json["length"] = length;
+ }
+ if (md5Checksum != null) {
+ _json["md5Checksum"] = md5Checksum;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (species != null) {
+ _json["species"] = species;
+ }
+ if (uri != null) {
+ _json["uri"] = uri;
+ }
+ return _json;
+ }
+}
+
+
+/** The callset search request. */
+class SearchCallsetsRequest {
+ /** The maximum number of callsets to return. */
+ core.String maxResults;
+
+ /**
+ * Only return callsets for which a substring of the name matches this string.
+ */
+ core.String name;
+
+ /**
+ * The continuation token, which is used to page through large result sets. To
+ * get the next page of results, set this parameter to the value of
+ * "nextPageToken" from the previous response.
+ */
+ core.String pageToken;
+
+ /**
+ * Restrict the query to callsets within the given variantsets. At least one
+ * ID must be provided.
+ */
+ core.List<core.String> variantsetIds;
+
+
+ SearchCallsetsRequest();
+
+ SearchCallsetsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("maxResults")) {
+ maxResults = _json["maxResults"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("pageToken")) {
+ pageToken = _json["pageToken"];
+ }
+ if (_json.containsKey("variantsetIds")) {
+ variantsetIds = _json["variantsetIds"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (maxResults != null) {
+ _json["maxResults"] = maxResults;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (pageToken != null) {
+ _json["pageToken"] = pageToken;
+ }
+ if (variantsetIds != null) {
+ _json["variantsetIds"] = variantsetIds;
+ }
+ return _json;
+ }
+}
+
+
+/** The callset search response. */
+class SearchCallsetsResponse {
+ /** The list of matching callsets. */
+ core.List<Callset> callsets;
+
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+
+ SearchCallsetsResponse();
+
+ SearchCallsetsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("callsets")) {
+ callsets = _json["callsets"].map((value) => new Callset.fromJson(value)).toList();
+ }
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (callsets != null) {
+ _json["callsets"] = callsets.map((value) => (value).toJson()).toList();
+ }
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ return _json;
+ }
+}
+
+
+/** The jobs search request. */
+class SearchJobsRequest {
+ /**
+ * If specified, only jobs created on or after this date, given in
+ * milliseconds since Unix epoch, will be returned.
+ */
+ core.String createdAfter;
+
+ /**
+ * If specified, only jobs created prior to this date, given in milliseconds
+ * since Unix epoch, will be returned.
+ */
+ core.String createdBefore;
+
+ /**
+ * Specifies the number of results to return in a single page. Defaults to
+ * 128. The maximum value is 256.
+ */
+ core.String maxResults;
+
+ /**
+ * The continuation token which is used to page through large result sets. To
+ * get the next page of results, set this parameter to the value of the
+ * "nextPageToken" from the previous response.
+ */
+ core.String pageToken;
+
+ /**
+ * Required. Only return jobs which belong to this Google Developers Console
+ * project. Only accepts project numbers.
+ */
+ core.String projectId;
+
+ /** Only return jobs which have a matching status. */
+ core.List<core.String> status;
+
+
+ SearchJobsRequest();
+
+ SearchJobsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("createdAfter")) {
+ createdAfter = _json["createdAfter"];
+ }
+ if (_json.containsKey("createdBefore")) {
+ createdBefore = _json["createdBefore"];
+ }
+ if (_json.containsKey("maxResults")) {
+ maxResults = _json["maxResults"];
+ }
+ if (_json.containsKey("pageToken")) {
+ pageToken = _json["pageToken"];
+ }
+ if (_json.containsKey("projectId")) {
+ projectId = _json["projectId"];
+ }
+ if (_json.containsKey("status")) {
+ status = _json["status"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (createdAfter != null) {
+ _json["createdAfter"] = createdAfter;
+ }
+ if (createdBefore != null) {
+ _json["createdBefore"] = createdBefore;
+ }
+ if (maxResults != null) {
+ _json["maxResults"] = maxResults;
+ }
+ if (pageToken != null) {
+ _json["pageToken"] = pageToken;
+ }
+ if (projectId != null) {
+ _json["projectId"] = projectId;
+ }
+ if (status != null) {
+ _json["status"] = status;
+ }
+ return _json;
+ }
+}
+
+
+/** The job search response. */
+class SearchJobsResponse {
+ /** The list of jobs results, ordered newest to oldest. */
+ core.List<Job> jobs;
+
+ /**
+ * The continuation token which is used to page through large result sets.
+ * Provide this value is a subsequent request to return the next page of
+ * results. This field will be empty if there are no more results.
+ */
+ core.String nextPageToken;
+
+
+ SearchJobsResponse();
+
+ SearchJobsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("jobs")) {
+ jobs = _json["jobs"].map((value) => new Job.fromJson(value)).toList();
+ }
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (jobs != null) {
+ _json["jobs"] = jobs.map((value) => (value).toJson()).toList();
+ }
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ return _json;
+ }
+}
+
+
+/** The read search request. */
+class SearchReadsRequest {
+ /**
+ * Specifies number of results to return in a single page. If unspecified, it
+ * will default to 256. The maximum value is 2048.
+ */
+ core.String maxResults;
+
+ /**
+ * The continuation token, which is used to page through large result sets. To
+ * get the next page of results, set this parameter to the value of
+ * "nextPageToken" from the previous response.
+ */
+ core.String pageToken;
+
+ /**
+ * The readsets within which to search for reads. At least one readset ID must
+ * be provided. All specified readsets must be aligned against a common set of
+ * reference sequences; this defines the genomic coordinates for the query.
+ */
+ core.List<core.String> readsetIds;
+
+ /**
+ * The end position (1-based, inclusive) of the target range. If specified,
+ * "sequenceName" must also be specified. Defaults to the end of the target
+ * reference sequence, if any.
+ */
+ core.String sequenceEnd;
+
+ /**
+ * Restricts the results to a particular reference sequence such as '1',
+ * 'chr1', or 'X'. The set of valid references sequences depends on the
+ * readsets specified. If set to "*", only unmapped Reads are returned.
+ */
+ core.String sequenceName;
+
+ /**
+ * The start position (1-based, inclusive) of the target range. If specified,
+ * "sequenceName" must also be specified. Defaults to the start of the target
+ * reference sequence, if any.
+ */
+ core.String sequenceStart;
+
+
+ SearchReadsRequest();
+
+ SearchReadsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("maxResults")) {
+ maxResults = _json["maxResults"];
+ }
+ if (_json.containsKey("pageToken")) {
+ pageToken = _json["pageToken"];
+ }
+ if (_json.containsKey("readsetIds")) {
+ readsetIds = _json["readsetIds"];
+ }
+ if (_json.containsKey("sequenceEnd")) {
+ sequenceEnd = _json["sequenceEnd"];
+ }
+ if (_json.containsKey("sequenceName")) {
+ sequenceName = _json["sequenceName"];
+ }
+ if (_json.containsKey("sequenceStart")) {
+ sequenceStart = _json["sequenceStart"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (maxResults != null) {
+ _json["maxResults"] = maxResults;
+ }
+ if (pageToken != null) {
+ _json["pageToken"] = pageToken;
+ }
+ if (readsetIds != null) {
+ _json["readsetIds"] = readsetIds;
+ }
+ if (sequenceEnd != null) {
+ _json["sequenceEnd"] = sequenceEnd;
+ }
+ if (sequenceName != null) {
+ _json["sequenceName"] = sequenceName;
+ }
+ if (sequenceStart != null) {
+ _json["sequenceStart"] = sequenceStart;
+ }
+ return _json;
+ }
+}
+
+
+/** The read search response. */
+class SearchReadsResponse {
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+ /**
+ * The list of matching Reads. The resulting Reads are sorted by position; the
+ * smallest positions are returned first. Unmapped reads, which have no
+ * position, are returned last and are further sorted alphabetically by name.
+ */
+ core.List<Read> reads;
+
+
+ SearchReadsResponse();
+
+ SearchReadsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ if (_json.containsKey("reads")) {
+ reads = _json["reads"].map((value) => new Read.fromJson(value)).toList();
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ if (reads != null) {
+ _json["reads"] = reads.map((value) => (value).toJson()).toList();
+ }
+ return _json;
+ }
+}
+
+
+/** The readset search request. */
+class SearchReadsetsRequest {
+ /**
+ * Restricts this query to readsets within the given datasets. At least one ID
+ * must be provided.
+ */
+ core.List<core.String> datasetIds;
+
+ /**
+ * Specifies number of results to return in a single page. If unspecified, it
+ * will default to 128. The maximum value is 1024.
+ */
+ core.String maxResults;
+
+ /**
+ * Only return readsets for which a substring of the name matches this string.
+ */
+ core.String name;
+
+ /**
+ * The continuation token, which is used to page through large result sets. To
+ * get the next page of results, set this parameter to the value of
+ * "nextPageToken" from the previous response.
+ */
+ core.String pageToken;
+
+
+ SearchReadsetsRequest();
+
+ SearchReadsetsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("datasetIds")) {
+ datasetIds = _json["datasetIds"];
+ }
+ if (_json.containsKey("maxResults")) {
+ maxResults = _json["maxResults"];
+ }
+ if (_json.containsKey("name")) {
+ name = _json["name"];
+ }
+ if (_json.containsKey("pageToken")) {
+ pageToken = _json["pageToken"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (datasetIds != null) {
+ _json["datasetIds"] = datasetIds;
+ }
+ if (maxResults != null) {
+ _json["maxResults"] = maxResults;
+ }
+ if (name != null) {
+ _json["name"] = name;
+ }
+ if (pageToken != null) {
+ _json["pageToken"] = pageToken;
+ }
+ return _json;
+ }
+}
+
+
+/** The readset search response. */
+class SearchReadsetsResponse {
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+ /** The list of matching Readsets. */
+ core.List<Readset> readsets;
+
+
+ SearchReadsetsResponse();
+
+ SearchReadsetsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ if (_json.containsKey("readsets")) {
+ readsets = _json["readsets"].map((value) => new Readset.fromJson(value)).toList();
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ if (readsets != null) {
+ _json["readsets"] = readsets.map((value) => (value).toJson()).toList();
+ }
+ return _json;
+ }
+}
+
+
+/** The variant search request. */
+class SearchVariantsRequest {
+ /**
+ * Only return variant calls which belong to callsets with these ids. Leaving
+ * this blank returns all variant calls. At most one of callsetNames or
+ * callsetIds should be provided.
+ */
+ core.List<core.String> callsetIds;
+
+ /**
+ * Only return variant calls which belong to callsets which have exactly these
+ * names. Leaving this blank returns all variant calls. At most one of
+ * callsetNames or callsetIds should be provided.
+ */
+ core.List<core.String> callsetNames;
+
+ /** Required. Only return variants on this contig. */
+ core.String contig;
+
+ /**
+ * Required. The end of the window (0-based, exclusive) for which overlapping
+ * variants should be returned.
+ */
+ core.String endPosition;
+
+ /** The maximum number of variants to return. */
+ core.String maxResults;
+
+ /**
+ * The continuation token, which is used to page through large result sets. To
+ * get the next page of results, set this parameter to the value of
+ * "nextPageToken" from the previous response.
+ */
+ core.String pageToken;
+
+ /**
+ * Required. The beginning of the window (0-based, inclusive) for which
+ * overlapping variants should be returned.
+ */
+ core.String startPosition;
+
+ /** Only return variants which have exactly this name. */
+ core.String variantName;
+
+ /** Required. The ID of the variantset to search. */
+ core.String variantsetId;
+
+
+ SearchVariantsRequest();
+
+ SearchVariantsRequest.fromJson(core.Map _json) {
+ if (_json.containsKey("callsetIds")) {
+ callsetIds = _json["callsetIds"];
+ }
+ if (_json.containsKey("callsetNames")) {
+ callsetNames = _json["callsetNames"];
+ }
+ if (_json.containsKey("contig")) {
+ contig = _json["contig"];
+ }
+ if (_json.containsKey("endPosition")) {
+ endPosition = _json["endPosition"];
+ }
+ if (_json.containsKey("maxResults")) {
+ maxResults = _json["maxResults"];
+ }
+ if (_json.containsKey("pageToken")) {
+ pageToken = _json["pageToken"];
+ }
+ if (_json.containsKey("startPosition")) {
+ startPosition = _json["startPosition"];
+ }
+ if (_json.containsKey("variantName")) {
+ variantName = _json["variantName"];
+ }
+ if (_json.containsKey("variantsetId")) {
+ variantsetId = _json["variantsetId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (callsetIds != null) {
+ _json["callsetIds"] = callsetIds;
+ }
+ if (callsetNames != null) {
+ _json["callsetNames"] = callsetNames;
+ }
+ if (contig != null) {
+ _json["contig"] = contig;
+ }
+ if (endPosition != null) {
+ _json["endPosition"] = endPosition;
+ }
+ if (maxResults != null) {
+ _json["maxResults"] = maxResults;
+ }
+ if (pageToken != null) {
+ _json["pageToken"] = pageToken;
+ }
+ if (startPosition != null) {
+ _json["startPosition"] = startPosition;
+ }
+ if (variantName != null) {
+ _json["variantName"] = variantName;
+ }
+ if (variantsetId != null) {
+ _json["variantsetId"] = variantsetId;
+ }
+ return _json;
+ }
+}
+
+
+/** The variant search response. */
+class SearchVariantsResponse {
+ /**
+ * The continuation token, which is used to page through large result sets.
+ * Provide this value in a subsequent request to return the next page of
+ * results. This field will be empty if there aren't any additional results.
+ */
+ core.String nextPageToken;
+
+ /** The list of matching Variants. */
+ core.List<Variant> variants;
+
+
+ SearchVariantsResponse();
+
+ SearchVariantsResponse.fromJson(core.Map _json) {
+ if (_json.containsKey("nextPageToken")) {
+ nextPageToken = _json["nextPageToken"];
+ }
+ if (_json.containsKey("variants")) {
+ variants = _json["variants"].map((value) => new Variant.fromJson(value)).toList();
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (nextPageToken != null) {
+ _json["nextPageToken"] = nextPageToken;
+ }
+ if (variants != null) {
+ _json["variants"] = variants.map((value) => (value).toJson()).toList();
+ }
+ return _json;
+ }
+}
+
+
+/**
+ * A Variant represents a change in DNA sequence relative to some reference. For
+ * example, a Variant could represent a SNP or an insertion. Variants belong to
+ * a Variantset.
+ */
+class Variant {
+ /** The bases that appear instead of the reference bases. */
+ core.List<core.String> alternateBases;
+
+ /**
+ * The variant calls for this particular variant. Each one represents the
+ * determination of genotype with respect to this variant.
+ */
+ core.List<Call> calls;
+
+ /** The contig on which this variant occurs. (such as 'chr20' or 'X') */
+ core.String contig;
+
+ /** The date this variant was created, in milliseconds from the epoch. */
+ core.String created;
+
+ /**
+ * The end position (0-based) of this variant. This corresponds to the first
+ * base after the last base in the reference allele. So, the length of the
+ * reference allele is (end - position). This is useful for variants that
+ * don't explicitly give alternate bases, for example large deletions.
+ */
+ core.String end;
+
+ /** The Google generated ID of the variant, immutable. */
+ core.String id;
+
+ /** A map of additional variant information. */
+ core.Map<core.String, core.List<core.String>> info;
+
+ /** Names for the variant, for example a RefSNP ID. */
+ core.List<core.String> names;
+
+ /**
+ * The position at which this variant occurs (0-based). This corresponds to
+ * the first base of the string of reference bases.
+ */
+ core.String position;
+
+ /**
+ * The reference bases for this variant. They start at the given position.
+ */
+ core.String referenceBases;
+
+ /** The ID of the variantset this variant belongs to. */
+ core.String variantsetId;
+
+
+ Variant();
+
+ Variant.fromJson(core.Map _json) {
+ if (_json.containsKey("alternateBases")) {
+ alternateBases = _json["alternateBases"];
+ }
+ if (_json.containsKey("calls")) {
+ calls = _json["calls"].map((value) => new Call.fromJson(value)).toList();
+ }
+ if (_json.containsKey("contig")) {
+ contig = _json["contig"];
+ }
+ if (_json.containsKey("created")) {
+ created = _json["created"];
+ }
+ if (_json.containsKey("end")) {
+ end = _json["end"];
+ }
+ if (_json.containsKey("id")) {
+ id = _json["id"];
+ }
+ if (_json.containsKey("info")) {
+ info = _json["info"];
+ }
+ if (_json.containsKey("names")) {
+ names = _json["names"];
+ }
+ if (_json.containsKey("position")) {
+ position = _json["position"];
+ }
+ if (_json.containsKey("referenceBases")) {
+ referenceBases = _json["referenceBases"];
+ }
+ if (_json.containsKey("variantsetId")) {
+ variantsetId = _json["variantsetId"];
+ }
+ }
+
+ core.Map toJson() {
+ var _json = new core.Map();
+ if (alternateBases != null) {
+ _json["alternateBases"] = alternateBases;
+ }
+ if (calls != null) {
+ _json["calls"] = calls.map((value) => (value).toJson()).toList();
+ }
+ if (contig != null) {
+ _json["contig"] = contig;
+ }
+ if (created != null) {
+ _json["created"] = created;
+ }
+ if (end != null) {
+ _json["end"] = end;
+ }
+ if (id != null) {
+ _json["id"] = id;
+ }
+ if (info != null) {
+ _json["info"] = info;
+ }
+ if (names != null) {
+ _json["names"] = names;
+ }
+ if (position != null) {
+ _json["position"] = position;
+ }
+ if (referenceBases != null) {
+ _json["referenceBases"] = referenceBases;
+ }
+ if (variantsetId != null) {
+ _json["variantsetId"] = variantsetId;
+ }
+ return _json;
+ }
+}
+
+

Powered by Google App Engine
This is Rietveld 408576698