| Index: generated/googleapis_beta/lib/genomics/v1beta.dart
|
| diff --git a/generated/googleapis_beta/lib/genomics/v1beta.dart b/generated/googleapis_beta/lib/genomics/v1beta.dart
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..c74ffdaaecfac47fa8e9f51d5e0e276505550ecd
|
| --- /dev/null
|
| +++ b/generated/googleapis_beta/lib/genomics/v1beta.dart
|
| @@ -0,0 +1,4284 @@
|
| +library googleapis_beta.genomics.v1beta;
|
| +
|
| +import "dart:core" as core;
|
| +import "dart:collection" as collection;
|
| +import "dart:async" as async;
|
| +import "dart:convert" as convert;
|
| +
|
| +import "package:crypto/crypto.dart" as crypto;
|
| +import 'package:http/http.dart' as http;
|
| +import '../src/common_internal.dart' as common_internal;
|
| +import '../common/common.dart' as common;
|
| +
|
| +export '../common/common.dart' show ApiRequestError;
|
| +export '../common/common.dart' show DetailedApiRequestError;
|
| +
|
| +/** Provides access to Genomics data. */
|
| +class GenomicsApi {
|
| + /** View and manage your data in Google BigQuery */
|
| + static const BigqueryScope = "https://www.googleapis.com/auth/bigquery";
|
| +
|
| + /** Manage your data in Google Cloud Storage */
|
| + static const DevstorageReadWriteScope = "https://www.googleapis.com/auth/devstorage.read_write";
|
| +
|
| + /** View and manage Genomics data */
|
| + static const GenomicsScope = "https://www.googleapis.com/auth/genomics";
|
| +
|
| + /** View Genomics data */
|
| + static const GenomicsReadonlyScope = "https://www.googleapis.com/auth/genomics.readonly";
|
| +
|
| +
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + BeaconsResourceApi get beacons => new BeaconsResourceApi(_requester);
|
| + CallsetsResourceApi get callsets => new CallsetsResourceApi(_requester);
|
| + DatasetsResourceApi get datasets => new DatasetsResourceApi(_requester);
|
| + ExperimentalResourceApi get experimental => new ExperimentalResourceApi(_requester);
|
| + JobsResourceApi get jobs => new JobsResourceApi(_requester);
|
| + ReadsResourceApi get reads => new ReadsResourceApi(_requester);
|
| + ReadsetsResourceApi get readsets => new ReadsetsResourceApi(_requester);
|
| + VariantsResourceApi get variants => new VariantsResourceApi(_requester);
|
| +
|
| + GenomicsApi(http.Client client) :
|
| + _requester = new common_internal.ApiRequester(client, "https://www.googleapis.com/", "/genomics/v1beta/");
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class BeaconsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + BeaconsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * This is an experimental API that provides a Global Alliance for Genomics
|
| + * and Health Beacon. It may change at any time.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantsetId] - The ID of the variantset to query over. It must be public.
|
| + * Private variantsets will return an unauthorized exception.
|
| + *
|
| + * [allele] - Required. The allele to look for ('A', 'C', 'G' or 'T').
|
| + *
|
| + * [contig] - Required. The contig to query over.
|
| + *
|
| + * [position] - Required. The 0-based position to query.
|
| + *
|
| + * Completes with a [Beacon].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Beacon> get(core.String variantsetId, {core.String allele, core.String contig, core.String position}) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (variantsetId == null) {
|
| + throw new core.ArgumentError("Parameter variantsetId is required.");
|
| + }
|
| + if (allele != null) {
|
| + _queryParams["allele"] = [allele];
|
| + }
|
| + if (contig != null) {
|
| + _queryParams["contig"] = [contig];
|
| + }
|
| + if (position != null) {
|
| + _queryParams["position"] = [position];
|
| + }
|
| +
|
| +
|
| + _url = 'beacons/' + common_internal.Escaper.ecapeVariable('$variantsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Beacon.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class CallsetsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + CallsetsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Creates a new callset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [Callset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Callset> create(Callset request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'callsets';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Callset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Deletes a callset.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [callsetId] - The ID of the callset to be deleted.
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future delete(core.String callsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (callsetId == null) {
|
| + throw new core.ArgumentError("Parameter callsetId is required.");
|
| + }
|
| +
|
| + _downloadOptions = null;
|
| +
|
| + _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "DELETE",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => null);
|
| + }
|
| +
|
| + /**
|
| + * Gets a callset by ID.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [callsetId] - The ID of the callset.
|
| + *
|
| + * Completes with a [Callset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Callset> get(core.String callsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (callsetId == null) {
|
| + throw new core.ArgumentError("Parameter callsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Callset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a callset. This method supports patch semantics.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [callsetId] - The ID of the callset to be updated.
|
| + *
|
| + * Completes with a [Callset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Callset> patch(Callset request, core.String callsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (callsetId == null) {
|
| + throw new core.ArgumentError("Parameter callsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PATCH",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Callset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a list of callsets matching the criteria.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [SearchCallsetsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<SearchCallsetsResponse> search(SearchCallsetsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'callsets/search';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new SearchCallsetsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a callset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [callsetId] - The ID of the callset to be updated.
|
| + *
|
| + * Completes with a [Callset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Callset> update(Callset request, core.String callsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (callsetId == null) {
|
| + throw new core.ArgumentError("Parameter callsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'callsets/' + common_internal.Escaper.ecapeVariable('$callsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PUT",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Callset.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class DatasetsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + DatasetsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Creates a new dataset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [Dataset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Dataset> create(Dataset request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'datasets';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Dataset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Deletes a dataset.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [datasetId] - The ID of the dataset to be deleted.
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future delete(core.String datasetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (datasetId == null) {
|
| + throw new core.ArgumentError("Parameter datasetId is required.");
|
| + }
|
| +
|
| + _downloadOptions = null;
|
| +
|
| + _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "DELETE",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => null);
|
| + }
|
| +
|
| + /**
|
| + * Gets a dataset by ID.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [datasetId] - The ID of the dataset.
|
| + *
|
| + * Completes with a [Dataset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Dataset> get(core.String datasetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (datasetId == null) {
|
| + throw new core.ArgumentError("Parameter datasetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Dataset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Lists all datasets.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [maxResults] - The maximum number of results returned by this request.
|
| + *
|
| + * [pageToken] - The continuation token, which is used to page through large
|
| + * result sets. To get the next page of results, set this parameter to the
|
| + * value of "nextPageToken" from the previous response.
|
| + *
|
| + * [projectId] - Only return datasets which belong to this Google Developers
|
| + * Console project. Only accepts project numbers.
|
| + *
|
| + * Completes with a [ListDatasetsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ListDatasetsResponse> list({core.String maxResults, core.String pageToken, core.String projectId}) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (maxResults != null) {
|
| + _queryParams["maxResults"] = [maxResults];
|
| + }
|
| + if (pageToken != null) {
|
| + _queryParams["pageToken"] = [pageToken];
|
| + }
|
| + if (projectId != null) {
|
| + _queryParams["projectId"] = [projectId];
|
| + }
|
| +
|
| +
|
| + _url = 'datasets';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ListDatasetsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a dataset. This method supports patch semantics.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [datasetId] - The ID of the dataset to be updated.
|
| + *
|
| + * Completes with a [Dataset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Dataset> patch(Dataset request, core.String datasetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (datasetId == null) {
|
| + throw new core.ArgumentError("Parameter datasetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PATCH",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Dataset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Undeletes a dataset by restoring a dataset which was deleted via this API.
|
| + * This operation is only possible for a week after the deletion occurred.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [datasetId] - The ID of the dataset to be undeleted.
|
| + *
|
| + * Completes with a [Dataset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Dataset> undelete(core.String datasetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (datasetId == null) {
|
| + throw new core.ArgumentError("Parameter datasetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId') + '/undelete';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Dataset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a dataset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [datasetId] - The ID of the dataset to be updated.
|
| + *
|
| + * Completes with a [Dataset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Dataset> update(Dataset request, core.String datasetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (datasetId == null) {
|
| + throw new core.ArgumentError("Parameter datasetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'datasets/' + common_internal.Escaper.ecapeVariable('$datasetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PUT",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Dataset.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ExperimentalResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + ExperimentalJobsResourceApi get jobs => new ExperimentalJobsResourceApi(_requester);
|
| +
|
| + ExperimentalResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ExperimentalJobsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + ExperimentalJobsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Creates and asynchronously runs an ad-hoc job. This is an experimental call
|
| + * and may be removed or changed at any time.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [ExperimentalCreateJobResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ExperimentalCreateJobResponse> create(ExperimentalCreateJobRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'experimental/jobs/create';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ExperimentalCreateJobResponse.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class JobsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + JobsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Cancels a job by ID. Note that it is possible for partial results to be
|
| + * generated and stored for cancelled jobs.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [jobId] - Required. The ID of the job.
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future cancel(core.String jobId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (jobId == null) {
|
| + throw new core.ArgumentError("Parameter jobId is required.");
|
| + }
|
| +
|
| + _downloadOptions = null;
|
| +
|
| + _url = 'jobs/' + common_internal.Escaper.ecapeVariable('$jobId') + '/cancel';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => null);
|
| + }
|
| +
|
| + /**
|
| + * Gets a job by ID.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [jobId] - Required. The ID of the job.
|
| + *
|
| + * Completes with a [Job].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Job> get(core.String jobId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (jobId == null) {
|
| + throw new core.ArgumentError("Parameter jobId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'jobs/' + common_internal.Escaper.ecapeVariable('$jobId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Job.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a list of jobs matching the criteria.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [SearchJobsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<SearchJobsResponse> search(SearchJobsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'jobs/search';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new SearchJobsResponse.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ReadsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + ReadsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Gets a list of reads for one or more readsets. Reads search operates over a
|
| + * genomic coordinate space of reference sequence & position defined over the
|
| + * reference sequences to which the requested readsets are aligned. If a
|
| + * target positional range is specified, search returns all reads whose
|
| + * alignment to the reference genome overlap the range. A query which
|
| + * specifies only readset IDs yields all reads in those readsets, including
|
| + * unmapped reads. All reads returned (including reads on subsequent pages)
|
| + * are ordered by genomic coordinate (reference sequence & position). Reads
|
| + * with equivalent genomic coordinates are returned in a deterministic order.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [SearchReadsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<SearchReadsResponse> search(SearchReadsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'reads/search';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new SearchReadsResponse.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ReadsetsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + ReadsetsCoveragebucketsResourceApi get coveragebuckets => new ReadsetsCoveragebucketsResourceApi(_requester);
|
| +
|
| + ReadsetsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Deletes a readset.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [readsetId] - The ID of the readset to be deleted. The caller must have
|
| + * WRITE permissions to the dataset associated with this readset.
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future delete(core.String readsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (readsetId == null) {
|
| + throw new core.ArgumentError("Parameter readsetId is required.");
|
| + }
|
| +
|
| + _downloadOptions = null;
|
| +
|
| + _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "DELETE",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => null);
|
| + }
|
| +
|
| + /**
|
| + * Exports readsets to a BAM file in Google Cloud Storage. Note that currently
|
| + * there may be some differences between exported BAM files and the original
|
| + * BAM file at the time of import. In particular, comments in the input file
|
| + * header will not be preserved, and some custom tags will be converted to
|
| + * strings.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [ExportReadsetsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ExportReadsetsResponse> export(ExportReadsetsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/export';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ExportReadsetsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a readset by ID.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [readsetId] - The ID of the readset.
|
| + *
|
| + * Completes with a [Readset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Readset> get(core.String readsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (readsetId == null) {
|
| + throw new core.ArgumentError("Parameter readsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Readset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Creates readsets by asynchronously importing the provided information. Note
|
| + * that currently comments in the input file header are not imported and some
|
| + * custom tags will be converted to strings, rather than preserving tag types.
|
| + * The caller must have WRITE permissions to the dataset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [ImportReadsetsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ImportReadsetsResponse> import(ImportReadsetsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/import';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ImportReadsetsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a readset. This method supports patch semantics.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [readsetId] - The ID of the readset to be updated. The caller must have
|
| + * WRITE permissions to the dataset associated with this readset.
|
| + *
|
| + * Completes with a [Readset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Readset> patch(Readset request, core.String readsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (readsetId == null) {
|
| + throw new core.ArgumentError("Parameter readsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PATCH",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Readset.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a list of readsets matching the criteria.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [SearchReadsetsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<SearchReadsetsResponse> search(SearchReadsetsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/search';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new SearchReadsetsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a readset.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [readsetId] - The ID of the readset to be updated. The caller must have
|
| + * WRITE permissions to the dataset associated with this readset.
|
| + *
|
| + * Completes with a [Readset].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Readset> update(Readset request, core.String readsetId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (readsetId == null) {
|
| + throw new core.ArgumentError("Parameter readsetId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PUT",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Readset.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ReadsetsCoveragebucketsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + ReadsetsCoveragebucketsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Lists fixed width coverage buckets for a readset, each of which correspond
|
| + * to a range of a reference sequence. Each bucket summarizes coverage
|
| + * information across its corresponding genomic range. Coverage is defined as
|
| + * the number of reads which are aligned to a given base in the reference
|
| + * sequence. Coverage buckets are available at various bucket widths, enabling
|
| + * various coverage "zoom levels". The caller must have READ permissions for
|
| + * the target readset.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [readsetId] - Required. The ID of the readset over which coverage is
|
| + * requested.
|
| + *
|
| + * [maxResults] - The maximum number of results to return in a single page. If
|
| + * unspecified, defaults to 1024. The maximum value is 2048.
|
| + *
|
| + * [pageToken] - The continuation token, which is used to page through large
|
| + * result sets. To get the next page of results, set this parameter to the
|
| + * value of "nextPageToken" from the previous response.
|
| + *
|
| + * [range_sequenceEnd] - The end position of the range on the reference,
|
| + * 1-based exclusive. If specified, sequenceName must also be specified.
|
| + *
|
| + * [range_sequenceName] - The reference sequence name, for example "chr1",
|
| + * "1", or "chrX".
|
| + *
|
| + * [range_sequenceStart] - The start position of the range on the reference,
|
| + * 1-based inclusive. If specified, sequenceName must also be specified.
|
| + *
|
| + * [targetBucketWidth] - The desired width of each reported coverage bucket in
|
| + * base pairs. This will be rounded down to the nearest precomputed bucket
|
| + * width; the value of which is returned as bucket_width in the response.
|
| + * Defaults to infinity (each bucket spans an entire reference sequence) or
|
| + * the length of the target range, if specified. The smallest precomputed
|
| + * bucket_width is currently 2048 base pairs; this is subject to change.
|
| + *
|
| + * Completes with a [ListCoverageBucketsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ListCoverageBucketsResponse> list(core.String readsetId, {core.String maxResults, core.String pageToken, core.String range_sequenceEnd, core.String range_sequenceName, core.String range_sequenceStart, core.String targetBucketWidth}) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (readsetId == null) {
|
| + throw new core.ArgumentError("Parameter readsetId is required.");
|
| + }
|
| + if (maxResults != null) {
|
| + _queryParams["maxResults"] = [maxResults];
|
| + }
|
| + if (pageToken != null) {
|
| + _queryParams["pageToken"] = [pageToken];
|
| + }
|
| + if (range_sequenceEnd != null) {
|
| + _queryParams["range.sequenceEnd"] = [range_sequenceEnd];
|
| + }
|
| + if (range_sequenceName != null) {
|
| + _queryParams["range.sequenceName"] = [range_sequenceName];
|
| + }
|
| + if (range_sequenceStart != null) {
|
| + _queryParams["range.sequenceStart"] = [range_sequenceStart];
|
| + }
|
| + if (targetBucketWidth != null) {
|
| + _queryParams["targetBucketWidth"] = [targetBucketWidth];
|
| + }
|
| +
|
| +
|
| + _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId') + '/coveragebuckets';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ListCoverageBucketsResponse.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class VariantsResourceApi {
|
| + final common_internal.ApiRequester _requester;
|
| +
|
| + VariantsResourceApi(common_internal.ApiRequester client) :
|
| + _requester = client;
|
| +
|
| + /**
|
| + * Creates a new variant.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [Variant].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Variant> create(Variant request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'variants';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Variant.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Deletes a variant.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantId] - The ID of the variant to be deleted.
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future delete(core.String variantId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (variantId == null) {
|
| + throw new core.ArgumentError("Parameter variantId is required.");
|
| + }
|
| +
|
| + _downloadOptions = null;
|
| +
|
| + _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "DELETE",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => null);
|
| + }
|
| +
|
| + /**
|
| + * Exports variant data to an external destination.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [ExportVariantsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ExportVariantsResponse> export(ExportVariantsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'variants/export';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ExportVariantsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a variant by ID.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantId] - The ID of the variant.
|
| + *
|
| + * Completes with a [Variant].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Variant> get(core.String variantId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (variantId == null) {
|
| + throw new core.ArgumentError("Parameter variantId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Variant.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a summary of all the variant data in a variantset.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantsetId] - Required. The ID of the variant set to get variant summary
|
| + * information for.
|
| + *
|
| + * Completes with a [GetVariantsSummaryResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<GetVariantsSummaryResponse> getSummary({core.String variantsetId}) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (variantsetId != null) {
|
| + _queryParams["variantsetId"] = [variantsetId];
|
| + }
|
| +
|
| +
|
| + _url = 'variants/summary';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "GET",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new GetVariantsSummaryResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Creates variant data by asynchronously importing the provided information.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [ImportVariantsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<ImportVariantsResponse> import(ImportVariantsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'variants/import';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new ImportVariantsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a variant. This method supports patch semantics.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantId] - The ID of the variant to be updated..
|
| + *
|
| + * Completes with a [Variant].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Variant> patch(Variant request, core.String variantId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (variantId == null) {
|
| + throw new core.ArgumentError("Parameter variantId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PATCH",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Variant.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Gets a list of variants matching the criteria.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * Completes with a [SearchVariantsResponse].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<SearchVariantsResponse> search(SearchVariantsRequest request) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| +
|
| +
|
| + _url = 'variants/search';
|
| +
|
| + var _response = _requester.request(_url,
|
| + "POST",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new SearchVariantsResponse.fromJson(data));
|
| + }
|
| +
|
| + /**
|
| + * Updates a variant.
|
| + *
|
| + * [request] - The metadata request object.
|
| + *
|
| + * Request parameters:
|
| + *
|
| + * [variantId] - The ID of the variant to be updated..
|
| + *
|
| + * Completes with a [Variant].
|
| + *
|
| + * Completes with a [common.ApiRequestError] if the API endpoint returned an
|
| + * error.
|
| + *
|
| + * If the used [http.Client] completes with an error when making a REST call,
|
| + * this method will complete with the same error.
|
| + */
|
| + async.Future<Variant> update(Variant request, core.String variantId) {
|
| + var _url = null;
|
| + var _queryParams = new core.Map();
|
| + var _uploadMedia = null;
|
| + var _uploadOptions = null;
|
| + var _downloadOptions = common.DownloadOptions.Metadata;
|
| + var _body = null;
|
| +
|
| + if (request != null) {
|
| + _body = convert.JSON.encode((request).toJson());
|
| + }
|
| + if (variantId == null) {
|
| + throw new core.ArgumentError("Parameter variantId is required.");
|
| + }
|
| +
|
| +
|
| + _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId');
|
| +
|
| + var _response = _requester.request(_url,
|
| + "PUT",
|
| + body: _body,
|
| + queryParams: _queryParams,
|
| + uploadOptions: _uploadOptions,
|
| + uploadMedia: _uploadMedia,
|
| + downloadOptions: _downloadOptions);
|
| + return _response.then((data) => new Variant.fromJson(data));
|
| + }
|
| +
|
| +}
|
| +
|
| +
|
| +
|
| +/**
|
| + * A beacon represents whether any variant call in a variantset has a specific
|
| + * allele at a particular position.
|
| + */
|
| +class Beacon {
|
| + /** True if the allele exists on any variant call, false otherwise. */
|
| + core.bool exists;
|
| +
|
| +
|
| + Beacon();
|
| +
|
| + Beacon.fromJson(core.Map _json) {
|
| + if (_json.containsKey("exists")) {
|
| + exists = _json["exists"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (exists != null) {
|
| + _json["exists"] = exists;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * A Call represents the determination of genotype with respect to a particular
|
| + * variant. It may include associated information such as quality and phasing.
|
| + * For example, a Call might assign a probability of 0.32 to the occurrence of a
|
| + * SNP named rs1234 in a callset with the name NA12345.
|
| + */
|
| +class Call {
|
| + /** The ID of the callset this variant call belongs to. */
|
| + core.String callsetId;
|
| +
|
| + /** The name of the callset this variant call belongs to. */
|
| + core.String callsetName;
|
| +
|
| + /**
|
| + * The genotype of this variant call. Each value represents either the value
|
| + * of the referenceBases field or a 1-based index into alternateBases. If a
|
| + * variant had a referenceBases field of "T" and an alternateBases value of
|
| + * ["A", "C"], and the genotype was [2, 1], that would mean the call
|
| + * represented the heterozygous value "CA" for this variant. If the genotype
|
| + * was instead [0, 1], the represented value would be "TA". Ordering of the
|
| + * genotype values is important if the phaseset field is present. If a
|
| + * genotype is not called (that is, a "." is present in the GT string) -1 is
|
| + * returned.
|
| + */
|
| + core.List<core.int> genotype;
|
| +
|
| + /**
|
| + * The genotype likelihoods for this variant call. Each array entry represents
|
| + * how likely a specific genotype is for this call. The value ordering is
|
| + * defined by the GL tag in the VCF spec.
|
| + */
|
| + core.List<core.double> genotypeLikelihood;
|
| +
|
| + /** A map of additional variant call information. */
|
| + core.Map<core.String, core.List<core.String>> info;
|
| +
|
| + /**
|
| + * If this field is present, this variant call's genotype ordering implies the
|
| + * phase of the bases and is consistent with any other variant calls on the
|
| + * same contig which have the same phaseset value. When importing data from
|
| + * VCF, if the genotype data was phased but no phase set was specified this
|
| + * field will be set to "*".
|
| + */
|
| + core.String phaseset;
|
| +
|
| +
|
| + Call();
|
| +
|
| + Call.fromJson(core.Map _json) {
|
| + if (_json.containsKey("callsetId")) {
|
| + callsetId = _json["callsetId"];
|
| + }
|
| + if (_json.containsKey("callsetName")) {
|
| + callsetName = _json["callsetName"];
|
| + }
|
| + if (_json.containsKey("genotype")) {
|
| + genotype = _json["genotype"];
|
| + }
|
| + if (_json.containsKey("genotypeLikelihood")) {
|
| + genotypeLikelihood = _json["genotypeLikelihood"];
|
| + }
|
| + if (_json.containsKey("info")) {
|
| + info = _json["info"];
|
| + }
|
| + if (_json.containsKey("phaseset")) {
|
| + phaseset = _json["phaseset"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (callsetId != null) {
|
| + _json["callsetId"] = callsetId;
|
| + }
|
| + if (callsetName != null) {
|
| + _json["callsetName"] = callsetName;
|
| + }
|
| + if (genotype != null) {
|
| + _json["genotype"] = genotype;
|
| + }
|
| + if (genotypeLikelihood != null) {
|
| + _json["genotypeLikelihood"] = genotypeLikelihood;
|
| + }
|
| + if (info != null) {
|
| + _json["info"] = info;
|
| + }
|
| + if (phaseset != null) {
|
| + _json["phaseset"] = phaseset;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** A Callset is a collection of Variant Calls. It belongs to a Variantset. */
|
| +class Callset {
|
| + /** The date this callset was created in milliseconds from the epoch. */
|
| + core.String created;
|
| +
|
| + /** The Google generated ID of the callset, immutable. */
|
| + core.String id;
|
| +
|
| + /** A map of additional callset information. */
|
| + core.Map<core.String, core.List<core.String>> info;
|
| +
|
| + /** The callset name. */
|
| + core.String name;
|
| +
|
| + /** The ID of the variantset this callset belongs to. */
|
| + core.String variantsetId;
|
| +
|
| +
|
| + Callset();
|
| +
|
| + Callset.fromJson(core.Map _json) {
|
| + if (_json.containsKey("created")) {
|
| + created = _json["created"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("info")) {
|
| + info = _json["info"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("variantsetId")) {
|
| + variantsetId = _json["variantsetId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (created != null) {
|
| + _json["created"] = created;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (info != null) {
|
| + _json["info"] = info;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (variantsetId != null) {
|
| + _json["variantsetId"] = variantsetId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * ContigBound records an upper bound for the starting coordinate of variants in
|
| + * a particular contig.
|
| + */
|
| +class ContigBound {
|
| + /** The contig the bound is associate with. */
|
| + core.String contig;
|
| +
|
| + /**
|
| + * An upper bound (inclusive) on the starting coordinate of any variant in the
|
| + * contig.
|
| + */
|
| + core.String upperBound;
|
| +
|
| +
|
| + ContigBound();
|
| +
|
| + ContigBound.fromJson(core.Map _json) {
|
| + if (_json.containsKey("contig")) {
|
| + contig = _json["contig"];
|
| + }
|
| + if (_json.containsKey("upperBound")) {
|
| + upperBound = _json["upperBound"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (contig != null) {
|
| + _json["contig"] = contig;
|
| + }
|
| + if (upperBound != null) {
|
| + _json["upperBound"] = upperBound;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * A bucket over which read coverage has been precomputed. A bucket corresponds
|
| + * to a specific range of the reference sequence.
|
| + */
|
| +class CoverageBucket {
|
| + /**
|
| + * The average number of reads which are aligned to each individual reference
|
| + * base in this bucket.
|
| + */
|
| + core.double meanCoverage;
|
| +
|
| + /** The genomic coordinate range spanned by this bucket. */
|
| + GenomicRange range;
|
| +
|
| +
|
| + CoverageBucket();
|
| +
|
| + CoverageBucket.fromJson(core.Map _json) {
|
| + if (_json.containsKey("meanCoverage")) {
|
| + meanCoverage = _json["meanCoverage"];
|
| + }
|
| + if (_json.containsKey("range")) {
|
| + range = new GenomicRange.fromJson(_json["range"]);
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (meanCoverage != null) {
|
| + _json["meanCoverage"] = meanCoverage;
|
| + }
|
| + if (range != null) {
|
| + _json["range"] = (range).toJson();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** A Dataset is a collection of genomic data. */
|
| +class Dataset {
|
| + /** The Google generated ID of the dataset, immutable. */
|
| + core.String id;
|
| +
|
| + /**
|
| + * Flag indicating whether or not a dataset is publicly viewable. If a dataset
|
| + * is not public, it inherits viewing permissions from its project.
|
| + */
|
| + core.bool isPublic;
|
| +
|
| + /** The dataset name. */
|
| + core.String name;
|
| +
|
| + /**
|
| + * The Google Developers Console project number that this dataset belongs to.
|
| + */
|
| + core.String projectId;
|
| +
|
| +
|
| + Dataset();
|
| +
|
| + Dataset.fromJson(core.Map _json) {
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("isPublic")) {
|
| + isPublic = _json["isPublic"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (isPublic != null) {
|
| + _json["isPublic"] = isPublic;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The job creation request. */
|
| +class ExperimentalCreateJobRequest {
|
| + /**
|
| + * Specifies whether or not to run the alignment pipeline. At least one of
|
| + * align or call_variants must be provided.
|
| + */
|
| + core.bool align;
|
| +
|
| + /**
|
| + * Specifies whether or not to run the variant calling pipeline. If specified,
|
| + * alignment will be performed first and the aligned BAMs will passed as input
|
| + * to the variant caller. At least one of align or call_variants must be
|
| + * provided.
|
| + */
|
| + core.bool callVariants;
|
| +
|
| + /**
|
| + * Specifies where to copy the results of certain pipelines. This shoud be in
|
| + * the form of "gs://bucket/path".
|
| + */
|
| + core.String gcsOutputPath;
|
| +
|
| + /**
|
| + * A list of Google Cloud Storage URIs of paired end .fastq files to operate
|
| + * upon. If specified, this represents the second file of each paired .fastq
|
| + * file. The first file of each pair should be specified in "sourceUris".
|
| + */
|
| + core.List<core.String> pairedSourceUris;
|
| +
|
| + /**
|
| + * Required. The Google Cloud Project ID with which to associate the request.
|
| + */
|
| + core.String projectId;
|
| +
|
| + /**
|
| + * A list of Google Cloud Storage URIs of data files to operate upon. These
|
| + * can be .bam, interleaved .fastq, or paired .fastq. If specifying paired
|
| + * .fastq files, the first of each pair of files should be listed here, and
|
| + * the second of each pair should be listed in "pairedSourceUris".
|
| + */
|
| + core.List<core.String> sourceUris;
|
| +
|
| +
|
| + ExperimentalCreateJobRequest();
|
| +
|
| + ExperimentalCreateJobRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("align")) {
|
| + align = _json["align"];
|
| + }
|
| + if (_json.containsKey("callVariants")) {
|
| + callVariants = _json["callVariants"];
|
| + }
|
| + if (_json.containsKey("gcsOutputPath")) {
|
| + gcsOutputPath = _json["gcsOutputPath"];
|
| + }
|
| + if (_json.containsKey("pairedSourceUris")) {
|
| + pairedSourceUris = _json["pairedSourceUris"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + if (_json.containsKey("sourceUris")) {
|
| + sourceUris = _json["sourceUris"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (align != null) {
|
| + _json["align"] = align;
|
| + }
|
| + if (callVariants != null) {
|
| + _json["callVariants"] = callVariants;
|
| + }
|
| + if (gcsOutputPath != null) {
|
| + _json["gcsOutputPath"] = gcsOutputPath;
|
| + }
|
| + if (pairedSourceUris != null) {
|
| + _json["pairedSourceUris"] = pairedSourceUris;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + if (sourceUris != null) {
|
| + _json["sourceUris"] = sourceUris;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The job creation response. */
|
| +class ExperimentalCreateJobResponse {
|
| + /** A job ID that can be used to get status information. */
|
| + core.String jobId;
|
| +
|
| +
|
| + ExperimentalCreateJobResponse();
|
| +
|
| + ExperimentalCreateJobResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobId")) {
|
| + jobId = _json["jobId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobId != null) {
|
| + _json["jobId"] = jobId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset export request. */
|
| +class ExportReadsetsRequest {
|
| + /**
|
| + * A Google Cloud Storage URI where the exported BAM file will be created. The
|
| + * currently authenticated user must have write access to the new file
|
| + * location. An error will be returned if the URI already contains data.
|
| + */
|
| + core.String exportUri;
|
| +
|
| + /**
|
| + * The Google Developers Console project number that owns this export. This is
|
| + * the project that will be billed.
|
| + */
|
| + core.String projectId;
|
| +
|
| + /** The IDs of the readsets to export. */
|
| + core.List<core.String> readsetIds;
|
| +
|
| + /**
|
| + * The reference names to export. If this is not specified, all reference
|
| + * sequences are exported. Use '*' to export unmapped reads.
|
| + */
|
| + core.List<core.String> referenceNames;
|
| +
|
| +
|
| + ExportReadsetsRequest();
|
| +
|
| + ExportReadsetsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("exportUri")) {
|
| + exportUri = _json["exportUri"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + if (_json.containsKey("readsetIds")) {
|
| + readsetIds = _json["readsetIds"];
|
| + }
|
| + if (_json.containsKey("referenceNames")) {
|
| + referenceNames = _json["referenceNames"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (exportUri != null) {
|
| + _json["exportUri"] = exportUri;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + if (readsetIds != null) {
|
| + _json["readsetIds"] = readsetIds;
|
| + }
|
| + if (referenceNames != null) {
|
| + _json["referenceNames"] = referenceNames;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset export response. */
|
| +class ExportReadsetsResponse {
|
| + /** A job ID that can be used to get status information. */
|
| + core.String jobId;
|
| +
|
| +
|
| + ExportReadsetsResponse();
|
| +
|
| + ExportReadsetsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobId")) {
|
| + jobId = _json["jobId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobId != null) {
|
| + _json["jobId"] = jobId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant data export request. */
|
| +class ExportVariantsRequest {
|
| + /**
|
| + * The BigQuery dataset to export data to. Note that this is distinct from the
|
| + * Genomics concept of "dataset". The caller must have WRITE access to this
|
| + * BigQuery dataset.
|
| + */
|
| + core.String bigqueryDataset;
|
| +
|
| + /**
|
| + * The BigQuery table to export data to. The caller must have WRITE access to
|
| + * this BigQuery table.
|
| + */
|
| + core.String bigqueryTable;
|
| +
|
| + /**
|
| + * If provided, only variant call information from the specified callsets will
|
| + * be exported. By default all variant calls are exported.
|
| + */
|
| + core.List<core.String> callsetIds;
|
| +
|
| + /**
|
| + * The format for the exported data.
|
| + * Possible string values are:
|
| + * - "bigquery"
|
| + */
|
| + core.String format;
|
| +
|
| + /**
|
| + * The Google Cloud project number that owns this export. This is the project
|
| + * that will be billed.
|
| + */
|
| + core.String projectId;
|
| +
|
| + /**
|
| + * Required. The ID of the variantset that contains variant data which should
|
| + * be exported. The caller must have READ access to this variantset.
|
| + */
|
| + core.String variantsetId;
|
| +
|
| +
|
| + ExportVariantsRequest();
|
| +
|
| + ExportVariantsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("bigqueryDataset")) {
|
| + bigqueryDataset = _json["bigqueryDataset"];
|
| + }
|
| + if (_json.containsKey("bigqueryTable")) {
|
| + bigqueryTable = _json["bigqueryTable"];
|
| + }
|
| + if (_json.containsKey("callsetIds")) {
|
| + callsetIds = _json["callsetIds"];
|
| + }
|
| + if (_json.containsKey("format")) {
|
| + format = _json["format"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + if (_json.containsKey("variantsetId")) {
|
| + variantsetId = _json["variantsetId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (bigqueryDataset != null) {
|
| + _json["bigqueryDataset"] = bigqueryDataset;
|
| + }
|
| + if (bigqueryTable != null) {
|
| + _json["bigqueryTable"] = bigqueryTable;
|
| + }
|
| + if (callsetIds != null) {
|
| + _json["callsetIds"] = callsetIds;
|
| + }
|
| + if (format != null) {
|
| + _json["format"] = format;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + if (variantsetId != null) {
|
| + _json["variantsetId"] = variantsetId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant data export response. */
|
| +class ExportVariantsResponse {
|
| + /** A job ID that can be used to get status information. */
|
| + core.String jobId;
|
| +
|
| +
|
| + ExportVariantsResponse();
|
| +
|
| + ExportVariantsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobId")) {
|
| + jobId = _json["jobId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobId != null) {
|
| + _json["jobId"] = jobId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * An inclusive, exclusive genomic coordinate range over a reference sequence.
|
| + */
|
| +class GenomicRange {
|
| + /**
|
| + * The end position of the range on the reference, 1-based exclusive. If
|
| + * specified, sequenceName must also be specified.
|
| + */
|
| + core.String sequenceEnd;
|
| +
|
| + /** The reference sequence name, for example "chr1", "1", or "chrX". */
|
| + core.String sequenceName;
|
| +
|
| + /**
|
| + * The start position of the range on the reference, 1-based inclusive. If
|
| + * specified, sequenceName must also be specified.
|
| + */
|
| + core.String sequenceStart;
|
| +
|
| +
|
| + GenomicRange();
|
| +
|
| + GenomicRange.fromJson(core.Map _json) {
|
| + if (_json.containsKey("sequenceEnd")) {
|
| + sequenceEnd = _json["sequenceEnd"];
|
| + }
|
| + if (_json.containsKey("sequenceName")) {
|
| + sequenceName = _json["sequenceName"];
|
| + }
|
| + if (_json.containsKey("sequenceStart")) {
|
| + sequenceStart = _json["sequenceStart"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (sequenceEnd != null) {
|
| + _json["sequenceEnd"] = sequenceEnd;
|
| + }
|
| + if (sequenceName != null) {
|
| + _json["sequenceName"] = sequenceName;
|
| + }
|
| + if (sequenceStart != null) {
|
| + _json["sequenceStart"] = sequenceStart;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variants summary response. */
|
| +class GetVariantsSummaryResponse {
|
| + /**
|
| + * A list of all contigs used by the variants in a dataset with associated
|
| + * coordinate upper bounds for each one.
|
| + */
|
| + core.List<ContigBound> contigBounds;
|
| +
|
| + /** The metadata associated with this dataset. */
|
| + core.List<Metadata> metadata;
|
| +
|
| +
|
| + GetVariantsSummaryResponse();
|
| +
|
| + GetVariantsSummaryResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("contigBounds")) {
|
| + contigBounds = _json["contigBounds"].map((value) => new ContigBound.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("metadata")) {
|
| + metadata = _json["metadata"].map((value) => new Metadata.fromJson(value)).toList();
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (contigBounds != null) {
|
| + _json["contigBounds"] = contigBounds.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (metadata != null) {
|
| + _json["metadata"] = metadata.map((value) => (value).toJson()).toList();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class Header {
|
| + /** (SO) Sorting order of alignments. */
|
| + core.String sortingOrder;
|
| +
|
| + /** (VN) BAM format version. */
|
| + core.String version;
|
| +
|
| +
|
| + Header();
|
| +
|
| + Header.fromJson(core.Map _json) {
|
| + if (_json.containsKey("sortingOrder")) {
|
| + sortingOrder = _json["sortingOrder"];
|
| + }
|
| + if (_json.containsKey("version")) {
|
| + version = _json["version"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (sortingOrder != null) {
|
| + _json["sortingOrder"] = sortingOrder;
|
| + }
|
| + if (version != null) {
|
| + _json["version"] = version;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The header section of the BAM/SAM file. */
|
| +class HeaderSection {
|
| + /** (@CO) One-line text comments. */
|
| + core.List<core.String> comments;
|
| +
|
| + /**
|
| + * [Deprecated] This field is deprecated and will no longer be populated.
|
| + * Please use filename instead.
|
| + */
|
| + core.String fileUri;
|
| +
|
| + /** The name of the file from which this data was imported. */
|
| + core.String filename;
|
| +
|
| + /** (@HD) The header line. */
|
| + core.List<Header> headers;
|
| +
|
| + /** (@PG) Programs. */
|
| + core.List<Program> programs;
|
| +
|
| + /** (@RG) Read group. */
|
| + core.List<ReadGroup> readGroups;
|
| +
|
| + /** (@SQ) Reference sequence dictionary. */
|
| + core.List<ReferenceSequence> refSequences;
|
| +
|
| +
|
| + HeaderSection();
|
| +
|
| + HeaderSection.fromJson(core.Map _json) {
|
| + if (_json.containsKey("comments")) {
|
| + comments = _json["comments"];
|
| + }
|
| + if (_json.containsKey("fileUri")) {
|
| + fileUri = _json["fileUri"];
|
| + }
|
| + if (_json.containsKey("filename")) {
|
| + filename = _json["filename"];
|
| + }
|
| + if (_json.containsKey("headers")) {
|
| + headers = _json["headers"].map((value) => new Header.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("programs")) {
|
| + programs = _json["programs"].map((value) => new Program.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("readGroups")) {
|
| + readGroups = _json["readGroups"].map((value) => new ReadGroup.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("refSequences")) {
|
| + refSequences = _json["refSequences"].map((value) => new ReferenceSequence.fromJson(value)).toList();
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (comments != null) {
|
| + _json["comments"] = comments;
|
| + }
|
| + if (fileUri != null) {
|
| + _json["fileUri"] = fileUri;
|
| + }
|
| + if (filename != null) {
|
| + _json["filename"] = filename;
|
| + }
|
| + if (headers != null) {
|
| + _json["headers"] = headers.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (programs != null) {
|
| + _json["programs"] = programs.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (readGroups != null) {
|
| + _json["readGroups"] = readGroups.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (refSequences != null) {
|
| + _json["refSequences"] = refSequences.map((value) => (value).toJson()).toList();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset import request. */
|
| +class ImportReadsetsRequest {
|
| + /**
|
| + * Required. The ID of the dataset these readsets will belong to. The caller
|
| + * must have WRITE permissions to this dataset.
|
| + */
|
| + core.String datasetId;
|
| +
|
| + /** A list of URIs pointing at BAM files in Google Cloud Storage. */
|
| + core.List<core.String> sourceUris;
|
| +
|
| +
|
| + ImportReadsetsRequest();
|
| +
|
| + ImportReadsetsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("datasetId")) {
|
| + datasetId = _json["datasetId"];
|
| + }
|
| + if (_json.containsKey("sourceUris")) {
|
| + sourceUris = _json["sourceUris"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (datasetId != null) {
|
| + _json["datasetId"] = datasetId;
|
| + }
|
| + if (sourceUris != null) {
|
| + _json["sourceUris"] = sourceUris;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset import response. */
|
| +class ImportReadsetsResponse {
|
| + /** A job ID that can be used to get status information. */
|
| + core.String jobId;
|
| +
|
| +
|
| + ImportReadsetsResponse();
|
| +
|
| + ImportReadsetsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobId")) {
|
| + jobId = _json["jobId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobId != null) {
|
| + _json["jobId"] = jobId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant data import request. */
|
| +class ImportVariantsRequest {
|
| + /**
|
| + * The format of the variant data being imported.
|
| + * Possible string values are:
|
| + * - "completeGenomics"
|
| + * - "vcf"
|
| + */
|
| + core.String format;
|
| +
|
| + /**
|
| + * A list of URIs pointing at VCF files in Google Cloud Storage. See the VCF
|
| + * Specification for more details on the input format.
|
| + */
|
| + core.List<core.String> sourceUris;
|
| +
|
| + /** Required. The variantset to which variant data should be imported. */
|
| + core.String variantsetId;
|
| +
|
| +
|
| + ImportVariantsRequest();
|
| +
|
| + ImportVariantsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("format")) {
|
| + format = _json["format"];
|
| + }
|
| + if (_json.containsKey("sourceUris")) {
|
| + sourceUris = _json["sourceUris"];
|
| + }
|
| + if (_json.containsKey("variantsetId")) {
|
| + variantsetId = _json["variantsetId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (format != null) {
|
| + _json["format"] = format;
|
| + }
|
| + if (sourceUris != null) {
|
| + _json["sourceUris"] = sourceUris;
|
| + }
|
| + if (variantsetId != null) {
|
| + _json["variantsetId"] = variantsetId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant data import response. */
|
| +class ImportVariantsResponse {
|
| + /** A job ID that can be used to get status information. */
|
| + core.String jobId;
|
| +
|
| +
|
| + ImportVariantsResponse();
|
| +
|
| + ImportVariantsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobId")) {
|
| + jobId = _json["jobId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobId != null) {
|
| + _json["jobId"] = jobId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * A Job represents an ongoing process that can be monitored for status
|
| + * information.
|
| + */
|
| +class Job {
|
| + /** The date this job was created, in milliseconds from the epoch. */
|
| + core.String created;
|
| +
|
| + /** A more detailed description of this job's current status. */
|
| + core.String description;
|
| +
|
| + /** Any errors that occurred during processing. */
|
| + core.List<core.String> errors;
|
| +
|
| + /** The job ID. */
|
| + core.String id;
|
| +
|
| + /**
|
| + * If this Job represents an import, this field will contain the IDs of the
|
| + * objects that were successfully imported.
|
| + */
|
| + core.List<core.String> importedIds;
|
| +
|
| + /**
|
| + * The Google Developers Console project number to which this job belongs.
|
| + */
|
| + core.String projectId;
|
| +
|
| + /** A summarized representation of the original service request. */
|
| + JobRequest request;
|
| +
|
| + /**
|
| + * The status of this job.
|
| + * Possible string values are:
|
| + * - "canceled"
|
| + * - "failure"
|
| + * - "new"
|
| + * - "pending"
|
| + * - "running"
|
| + * - "success"
|
| + * - "unknownStatus"
|
| + */
|
| + core.String status;
|
| +
|
| + /** Any warnings that occurred during processing. */
|
| + core.List<core.String> warnings;
|
| +
|
| +
|
| + Job();
|
| +
|
| + Job.fromJson(core.Map _json) {
|
| + if (_json.containsKey("created")) {
|
| + created = _json["created"];
|
| + }
|
| + if (_json.containsKey("description")) {
|
| + description = _json["description"];
|
| + }
|
| + if (_json.containsKey("errors")) {
|
| + errors = _json["errors"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("importedIds")) {
|
| + importedIds = _json["importedIds"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + if (_json.containsKey("request")) {
|
| + request = new JobRequest.fromJson(_json["request"]);
|
| + }
|
| + if (_json.containsKey("status")) {
|
| + status = _json["status"];
|
| + }
|
| + if (_json.containsKey("warnings")) {
|
| + warnings = _json["warnings"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (created != null) {
|
| + _json["created"] = created;
|
| + }
|
| + if (description != null) {
|
| + _json["description"] = description;
|
| + }
|
| + if (errors != null) {
|
| + _json["errors"] = errors;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (importedIds != null) {
|
| + _json["importedIds"] = importedIds;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + if (request != null) {
|
| + _json["request"] = (request).toJson();
|
| + }
|
| + if (status != null) {
|
| + _json["status"] = status;
|
| + }
|
| + if (warnings != null) {
|
| + _json["warnings"] = warnings;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** A summary representation of the service request that spawned the job. */
|
| +class JobRequest {
|
| + /**
|
| + * The data destination of the request, for example, a Google BigQuery Table
|
| + * or Dataset ID.
|
| + */
|
| + core.List<core.String> destination;
|
| +
|
| + /**
|
| + * The data source of the request, for example, a Google Cloud Storage object
|
| + * path or Readset ID.
|
| + */
|
| + core.List<core.String> source;
|
| +
|
| + /**
|
| + * The original request type.
|
| + * Possible string values are:
|
| + * - "experimentalCreateJob"
|
| + * - "exportReadsets"
|
| + * - "exportVariants"
|
| + * - "importReadsets"
|
| + * - "importVariants"
|
| + * - "unknownType"
|
| + */
|
| + core.String type;
|
| +
|
| +
|
| + JobRequest();
|
| +
|
| + JobRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("destination")) {
|
| + destination = _json["destination"];
|
| + }
|
| + if (_json.containsKey("source")) {
|
| + source = _json["source"];
|
| + }
|
| + if (_json.containsKey("type")) {
|
| + type = _json["type"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (destination != null) {
|
| + _json["destination"] = destination;
|
| + }
|
| + if (source != null) {
|
| + _json["source"] = source;
|
| + }
|
| + if (type != null) {
|
| + _json["type"] = type;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ListCoverageBucketsResponse {
|
| + /**
|
| + * The length of each coverage bucket in base pairs. Note that buckets at the
|
| + * end of a reference sequence may be shorter. This value is omitted if the
|
| + * bucket width is infinity (the default behaviour, with no range or
|
| + * target_bucket_width).
|
| + */
|
| + core.String bucketWidth;
|
| +
|
| + /**
|
| + * The coverage buckets. The list of buckets is sparse; a bucket with 0
|
| + * overlapping reads is not returned. A bucket never crosses more than one
|
| + * reference sequence. Each bucket has width bucket_width, unless its end is
|
| + * is the end of the reference sequence.
|
| + */
|
| + core.List<CoverageBucket> coverageBuckets;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| +
|
| + ListCoverageBucketsResponse();
|
| +
|
| + ListCoverageBucketsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("bucketWidth")) {
|
| + bucketWidth = _json["bucketWidth"];
|
| + }
|
| + if (_json.containsKey("coverageBuckets")) {
|
| + coverageBuckets = _json["coverageBuckets"].map((value) => new CoverageBucket.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (bucketWidth != null) {
|
| + _json["bucketWidth"] = bucketWidth;
|
| + }
|
| + if (coverageBuckets != null) {
|
| + _json["coverageBuckets"] = coverageBuckets.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The dataset list response. */
|
| +class ListDatasetsResponse {
|
| + /** The list of matching Datasets. */
|
| + core.List<Dataset> datasets;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| +
|
| + ListDatasetsResponse();
|
| +
|
| + ListDatasetsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("datasets")) {
|
| + datasets = _json["datasets"].map((value) => new Dataset.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (datasets != null) {
|
| + _json["datasets"] = datasets.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * Metadata describes a single piece of variant call metadata. These data
|
| + * include a top level key and either a single value string (value) or a list of
|
| + * key-value pairs (info.) Value and info are mutually exclusive.
|
| + */
|
| +class Metadata {
|
| + /** A textual description of this metadata. */
|
| + core.String description;
|
| +
|
| + /**
|
| + * User-provided ID field, not enforced by this API. Two or more pieces of
|
| + * structured metadata with identical id and key fields are considered
|
| + * equivalent.
|
| + */
|
| + core.String id;
|
| +
|
| + /** Remaining structured metadata key-value pairs. */
|
| + core.Map<core.String, core.List<core.String>> info;
|
| +
|
| + /** The top-level key. */
|
| + core.String key;
|
| +
|
| + /**
|
| + * The number of values that can be included in a field described by this
|
| + * metadata.
|
| + */
|
| + core.String number;
|
| +
|
| + /**
|
| + * The type of data. Possible types include: Integer, Float, Flag, Character,
|
| + * and String.
|
| + * Possible string values are:
|
| + * - "character"
|
| + * - "flag"
|
| + * - "float"
|
| + * - "integer"
|
| + * - "string"
|
| + * - "unknownType"
|
| + */
|
| + core.String type;
|
| +
|
| + /** The value field for simple metadata */
|
| + core.String value;
|
| +
|
| +
|
| + Metadata();
|
| +
|
| + Metadata.fromJson(core.Map _json) {
|
| + if (_json.containsKey("description")) {
|
| + description = _json["description"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("info")) {
|
| + info = _json["info"];
|
| + }
|
| + if (_json.containsKey("key")) {
|
| + key = _json["key"];
|
| + }
|
| + if (_json.containsKey("number")) {
|
| + number = _json["number"];
|
| + }
|
| + if (_json.containsKey("type")) {
|
| + type = _json["type"];
|
| + }
|
| + if (_json.containsKey("value")) {
|
| + value = _json["value"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (description != null) {
|
| + _json["description"] = description;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (info != null) {
|
| + _json["info"] = info;
|
| + }
|
| + if (key != null) {
|
| + _json["key"] = key;
|
| + }
|
| + if (number != null) {
|
| + _json["number"] = number;
|
| + }
|
| + if (type != null) {
|
| + _json["type"] = type;
|
| + }
|
| + if (value != null) {
|
| + _json["value"] = value;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class Program {
|
| + /** (CL) Command line. */
|
| + core.String commandLine;
|
| +
|
| + /** (ID) Program record identifier. */
|
| + core.String id;
|
| +
|
| + /** (PN) Program name. */
|
| + core.String name;
|
| +
|
| + /** (PP) Previous program ID. */
|
| + core.String prevProgramId;
|
| +
|
| + /** (VN) Program version. */
|
| + core.String version;
|
| +
|
| +
|
| + Program();
|
| +
|
| + Program.fromJson(core.Map _json) {
|
| + if (_json.containsKey("commandLine")) {
|
| + commandLine = _json["commandLine"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("prevProgramId")) {
|
| + prevProgramId = _json["prevProgramId"];
|
| + }
|
| + if (_json.containsKey("version")) {
|
| + version = _json["version"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (commandLine != null) {
|
| + _json["commandLine"] = commandLine;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (prevProgramId != null) {
|
| + _json["prevProgramId"] = prevProgramId;
|
| + }
|
| + if (version != null) {
|
| + _json["version"] = version;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * A Read is a group of bases that may or may not have been aligned to a
|
| + * reference. It contains quality information and other metadata.
|
| + */
|
| +class Read {
|
| + /**
|
| + * The originalBases after the cigar field has been applied. Deletions are
|
| + * represented with '-' and insertions are omitted.
|
| + */
|
| + core.String alignedBases;
|
| +
|
| + /**
|
| + * Represents the quality of each base in this read. Each character represents
|
| + * one base. To get the quality, take the ASCII value of the character and
|
| + * subtract 33. (QUAL)
|
| + */
|
| + core.String baseQuality;
|
| +
|
| + /**
|
| + * A condensed representation of how this read matches up to the reference.
|
| + * (CIGAR)
|
| + */
|
| + core.String cigar;
|
| +
|
| + /**
|
| + * Each bit of this number has a different meaning if enabled. See the full
|
| + * BAM spec for more details. (FLAG)
|
| + */
|
| + core.int flags;
|
| +
|
| + /** The Google generated ID of the read, immutable. */
|
| + core.String id;
|
| +
|
| + /**
|
| + * A score up to 255 that represents how likely this read's aligned position
|
| + * is to be correct. A higher value is better. (MAPQ)
|
| + */
|
| + core.int mappingQuality;
|
| +
|
| + /** The 1-based start position of the paired read. (PNEXT) */
|
| + core.int matePosition;
|
| +
|
| + /**
|
| + * The name of the sequence that the paired read is aligned to. This is
|
| + * usually the same as referenceSequenceName. (RNEXT)
|
| + */
|
| + core.String mateReferenceSequenceName;
|
| +
|
| + /**
|
| + * The name of the read. When imported from a BAM file, this is the query
|
| + * template name. (QNAME)
|
| + */
|
| + core.String name;
|
| +
|
| + /** The list of bases that this read represents (such as "CATCGA"). (SEQ) */
|
| + core.String originalBases;
|
| +
|
| + /**
|
| + * The 1-based start position of the aligned read. If the first base starts at
|
| + * the very beginning of the reference sequence, then the position would be
|
| + * '1'. (POS)
|
| + */
|
| + core.int position;
|
| +
|
| + /** The ID of the readset this read belongs to. */
|
| + core.String readsetId;
|
| +
|
| + /**
|
| + * The name of the sequence that this read is aligned to. This would be, for
|
| + * example, 'X' for the X Chromosome or '20' for Chromosome 20. (RNAME)
|
| + */
|
| + core.String referenceSequenceName;
|
| +
|
| + /** A map of additional read information. (TAG) */
|
| + core.Map<core.String, core.List<core.String>> tags;
|
| +
|
| + /**
|
| + * Length of the original piece of DNA that produced both this read and the
|
| + * paired read. (TLEN)
|
| + */
|
| + core.int templateLength;
|
| +
|
| +
|
| + Read();
|
| +
|
| + Read.fromJson(core.Map _json) {
|
| + if (_json.containsKey("alignedBases")) {
|
| + alignedBases = _json["alignedBases"];
|
| + }
|
| + if (_json.containsKey("baseQuality")) {
|
| + baseQuality = _json["baseQuality"];
|
| + }
|
| + if (_json.containsKey("cigar")) {
|
| + cigar = _json["cigar"];
|
| + }
|
| + if (_json.containsKey("flags")) {
|
| + flags = _json["flags"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("mappingQuality")) {
|
| + mappingQuality = _json["mappingQuality"];
|
| + }
|
| + if (_json.containsKey("matePosition")) {
|
| + matePosition = _json["matePosition"];
|
| + }
|
| + if (_json.containsKey("mateReferenceSequenceName")) {
|
| + mateReferenceSequenceName = _json["mateReferenceSequenceName"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("originalBases")) {
|
| + originalBases = _json["originalBases"];
|
| + }
|
| + if (_json.containsKey("position")) {
|
| + position = _json["position"];
|
| + }
|
| + if (_json.containsKey("readsetId")) {
|
| + readsetId = _json["readsetId"];
|
| + }
|
| + if (_json.containsKey("referenceSequenceName")) {
|
| + referenceSequenceName = _json["referenceSequenceName"];
|
| + }
|
| + if (_json.containsKey("tags")) {
|
| + tags = _json["tags"];
|
| + }
|
| + if (_json.containsKey("templateLength")) {
|
| + templateLength = _json["templateLength"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (alignedBases != null) {
|
| + _json["alignedBases"] = alignedBases;
|
| + }
|
| + if (baseQuality != null) {
|
| + _json["baseQuality"] = baseQuality;
|
| + }
|
| + if (cigar != null) {
|
| + _json["cigar"] = cigar;
|
| + }
|
| + if (flags != null) {
|
| + _json["flags"] = flags;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (mappingQuality != null) {
|
| + _json["mappingQuality"] = mappingQuality;
|
| + }
|
| + if (matePosition != null) {
|
| + _json["matePosition"] = matePosition;
|
| + }
|
| + if (mateReferenceSequenceName != null) {
|
| + _json["mateReferenceSequenceName"] = mateReferenceSequenceName;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (originalBases != null) {
|
| + _json["originalBases"] = originalBases;
|
| + }
|
| + if (position != null) {
|
| + _json["position"] = position;
|
| + }
|
| + if (readsetId != null) {
|
| + _json["readsetId"] = readsetId;
|
| + }
|
| + if (referenceSequenceName != null) {
|
| + _json["referenceSequenceName"] = referenceSequenceName;
|
| + }
|
| + if (tags != null) {
|
| + _json["tags"] = tags;
|
| + }
|
| + if (templateLength != null) {
|
| + _json["templateLength"] = templateLength;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ReadGroup {
|
| + /** (DT) Date the run was produced (ISO8601 date or date/time). */
|
| + core.String date;
|
| +
|
| + /** (DS) Description. */
|
| + core.String description;
|
| +
|
| + /**
|
| + * (FO) Flow order. The array of nucleotide bases that correspond to the
|
| + * nucleotides used for each flow of each read.
|
| + */
|
| + core.String flowOrder;
|
| +
|
| + /** (ID) Read group identifier. */
|
| + core.String id;
|
| +
|
| + /**
|
| + * (KS) The array of nucleotide bases that correspond to the key sequence of
|
| + * each read.
|
| + */
|
| + core.String keySequence;
|
| +
|
| + /** (LS) Library. */
|
| + core.String library;
|
| +
|
| + /** (PU) Platform unit. */
|
| + core.String platformUnit;
|
| +
|
| + /** (PI) Predicted median insert size. */
|
| + core.int predictedInsertSize;
|
| +
|
| + /** (PG) Programs used for processing the read group. */
|
| + core.String processingProgram;
|
| +
|
| + /** (SM) Sample. */
|
| + core.String sample;
|
| +
|
| + /** (CN) Name of sequencing center producing the read. */
|
| + core.String sequencingCenterName;
|
| +
|
| + /** (PL) Platform/technology used to produce the reads. */
|
| + core.String sequencingTechnology;
|
| +
|
| +
|
| + ReadGroup();
|
| +
|
| + ReadGroup.fromJson(core.Map _json) {
|
| + if (_json.containsKey("date")) {
|
| + date = _json["date"];
|
| + }
|
| + if (_json.containsKey("description")) {
|
| + description = _json["description"];
|
| + }
|
| + if (_json.containsKey("flowOrder")) {
|
| + flowOrder = _json["flowOrder"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("keySequence")) {
|
| + keySequence = _json["keySequence"];
|
| + }
|
| + if (_json.containsKey("library")) {
|
| + library = _json["library"];
|
| + }
|
| + if (_json.containsKey("platformUnit")) {
|
| + platformUnit = _json["platformUnit"];
|
| + }
|
| + if (_json.containsKey("predictedInsertSize")) {
|
| + predictedInsertSize = _json["predictedInsertSize"];
|
| + }
|
| + if (_json.containsKey("processingProgram")) {
|
| + processingProgram = _json["processingProgram"];
|
| + }
|
| + if (_json.containsKey("sample")) {
|
| + sample = _json["sample"];
|
| + }
|
| + if (_json.containsKey("sequencingCenterName")) {
|
| + sequencingCenterName = _json["sequencingCenterName"];
|
| + }
|
| + if (_json.containsKey("sequencingTechnology")) {
|
| + sequencingTechnology = _json["sequencingTechnology"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (date != null) {
|
| + _json["date"] = date;
|
| + }
|
| + if (description != null) {
|
| + _json["description"] = description;
|
| + }
|
| + if (flowOrder != null) {
|
| + _json["flowOrder"] = flowOrder;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (keySequence != null) {
|
| + _json["keySequence"] = keySequence;
|
| + }
|
| + if (library != null) {
|
| + _json["library"] = library;
|
| + }
|
| + if (platformUnit != null) {
|
| + _json["platformUnit"] = platformUnit;
|
| + }
|
| + if (predictedInsertSize != null) {
|
| + _json["predictedInsertSize"] = predictedInsertSize;
|
| + }
|
| + if (processingProgram != null) {
|
| + _json["processingProgram"] = processingProgram;
|
| + }
|
| + if (sample != null) {
|
| + _json["sample"] = sample;
|
| + }
|
| + if (sequencingCenterName != null) {
|
| + _json["sequencingCenterName"] = sequencingCenterName;
|
| + }
|
| + if (sequencingTechnology != null) {
|
| + _json["sequencingTechnology"] = sequencingTechnology;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** A Readset is a collection of Reads. */
|
| +class Readset {
|
| + /** The ID of the dataset this readset belongs to. */
|
| + core.String datasetId;
|
| +
|
| + /**
|
| + * File information from the original BAM import. See the BAM format
|
| + * specification for additional information on each field.
|
| + */
|
| + core.List<HeaderSection> fileData;
|
| +
|
| + /** The Google generated ID of the readset, immutable. */
|
| + core.String id;
|
| +
|
| + /** The readset name, typically this is the sample name. */
|
| + core.String name;
|
| +
|
| +
|
| + Readset();
|
| +
|
| + Readset.fromJson(core.Map _json) {
|
| + if (_json.containsKey("datasetId")) {
|
| + datasetId = _json["datasetId"];
|
| + }
|
| + if (_json.containsKey("fileData")) {
|
| + fileData = _json["fileData"].map((value) => new HeaderSection.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (datasetId != null) {
|
| + _json["datasetId"] = datasetId;
|
| + }
|
| + if (fileData != null) {
|
| + _json["fileData"] = fileData.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** Not documented yet. */
|
| +class ReferenceSequence {
|
| + /** (AS) Genome assembly identifier. */
|
| + core.String assemblyId;
|
| +
|
| + /** (LN) Reference sequence length. */
|
| + core.int length;
|
| +
|
| + /**
|
| + * (M5) MD5 checksum of the sequence in the uppercase, excluding spaces but
|
| + * including pads as *.
|
| + */
|
| + core.String md5Checksum;
|
| +
|
| + /** (SN) Reference sequence name. */
|
| + core.String name;
|
| +
|
| + /** (SP) Species. */
|
| + core.String species;
|
| +
|
| + /** (UR) URI of the sequence. */
|
| + core.String uri;
|
| +
|
| +
|
| + ReferenceSequence();
|
| +
|
| + ReferenceSequence.fromJson(core.Map _json) {
|
| + if (_json.containsKey("assemblyId")) {
|
| + assemblyId = _json["assemblyId"];
|
| + }
|
| + if (_json.containsKey("length")) {
|
| + length = _json["length"];
|
| + }
|
| + if (_json.containsKey("md5Checksum")) {
|
| + md5Checksum = _json["md5Checksum"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("species")) {
|
| + species = _json["species"];
|
| + }
|
| + if (_json.containsKey("uri")) {
|
| + uri = _json["uri"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (assemblyId != null) {
|
| + _json["assemblyId"] = assemblyId;
|
| + }
|
| + if (length != null) {
|
| + _json["length"] = length;
|
| + }
|
| + if (md5Checksum != null) {
|
| + _json["md5Checksum"] = md5Checksum;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (species != null) {
|
| + _json["species"] = species;
|
| + }
|
| + if (uri != null) {
|
| + _json["uri"] = uri;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The callset search request. */
|
| +class SearchCallsetsRequest {
|
| + /** The maximum number of callsets to return. */
|
| + core.String maxResults;
|
| +
|
| + /**
|
| + * Only return callsets for which a substring of the name matches this string.
|
| + */
|
| + core.String name;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets. To
|
| + * get the next page of results, set this parameter to the value of
|
| + * "nextPageToken" from the previous response.
|
| + */
|
| + core.String pageToken;
|
| +
|
| + /**
|
| + * Restrict the query to callsets within the given variantsets. At least one
|
| + * ID must be provided.
|
| + */
|
| + core.List<core.String> variantsetIds;
|
| +
|
| +
|
| + SearchCallsetsRequest();
|
| +
|
| + SearchCallsetsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("maxResults")) {
|
| + maxResults = _json["maxResults"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("pageToken")) {
|
| + pageToken = _json["pageToken"];
|
| + }
|
| + if (_json.containsKey("variantsetIds")) {
|
| + variantsetIds = _json["variantsetIds"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (maxResults != null) {
|
| + _json["maxResults"] = maxResults;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (pageToken != null) {
|
| + _json["pageToken"] = pageToken;
|
| + }
|
| + if (variantsetIds != null) {
|
| + _json["variantsetIds"] = variantsetIds;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The callset search response. */
|
| +class SearchCallsetsResponse {
|
| + /** The list of matching callsets. */
|
| + core.List<Callset> callsets;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| +
|
| + SearchCallsetsResponse();
|
| +
|
| + SearchCallsetsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("callsets")) {
|
| + callsets = _json["callsets"].map((value) => new Callset.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (callsets != null) {
|
| + _json["callsets"] = callsets.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The jobs search request. */
|
| +class SearchJobsRequest {
|
| + /**
|
| + * If specified, only jobs created on or after this date, given in
|
| + * milliseconds since Unix epoch, will be returned.
|
| + */
|
| + core.String createdAfter;
|
| +
|
| + /**
|
| + * If specified, only jobs created prior to this date, given in milliseconds
|
| + * since Unix epoch, will be returned.
|
| + */
|
| + core.String createdBefore;
|
| +
|
| + /**
|
| + * Specifies the number of results to return in a single page. Defaults to
|
| + * 128. The maximum value is 256.
|
| + */
|
| + core.String maxResults;
|
| +
|
| + /**
|
| + * The continuation token which is used to page through large result sets. To
|
| + * get the next page of results, set this parameter to the value of the
|
| + * "nextPageToken" from the previous response.
|
| + */
|
| + core.String pageToken;
|
| +
|
| + /**
|
| + * Required. Only return jobs which belong to this Google Developers Console
|
| + * project. Only accepts project numbers.
|
| + */
|
| + core.String projectId;
|
| +
|
| + /** Only return jobs which have a matching status. */
|
| + core.List<core.String> status;
|
| +
|
| +
|
| + SearchJobsRequest();
|
| +
|
| + SearchJobsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("createdAfter")) {
|
| + createdAfter = _json["createdAfter"];
|
| + }
|
| + if (_json.containsKey("createdBefore")) {
|
| + createdBefore = _json["createdBefore"];
|
| + }
|
| + if (_json.containsKey("maxResults")) {
|
| + maxResults = _json["maxResults"];
|
| + }
|
| + if (_json.containsKey("pageToken")) {
|
| + pageToken = _json["pageToken"];
|
| + }
|
| + if (_json.containsKey("projectId")) {
|
| + projectId = _json["projectId"];
|
| + }
|
| + if (_json.containsKey("status")) {
|
| + status = _json["status"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (createdAfter != null) {
|
| + _json["createdAfter"] = createdAfter;
|
| + }
|
| + if (createdBefore != null) {
|
| + _json["createdBefore"] = createdBefore;
|
| + }
|
| + if (maxResults != null) {
|
| + _json["maxResults"] = maxResults;
|
| + }
|
| + if (pageToken != null) {
|
| + _json["pageToken"] = pageToken;
|
| + }
|
| + if (projectId != null) {
|
| + _json["projectId"] = projectId;
|
| + }
|
| + if (status != null) {
|
| + _json["status"] = status;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The job search response. */
|
| +class SearchJobsResponse {
|
| + /** The list of jobs results, ordered newest to oldest. */
|
| + core.List<Job> jobs;
|
| +
|
| + /**
|
| + * The continuation token which is used to page through large result sets.
|
| + * Provide this value is a subsequent request to return the next page of
|
| + * results. This field will be empty if there are no more results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| +
|
| + SearchJobsResponse();
|
| +
|
| + SearchJobsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("jobs")) {
|
| + jobs = _json["jobs"].map((value) => new Job.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (jobs != null) {
|
| + _json["jobs"] = jobs.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The read search request. */
|
| +class SearchReadsRequest {
|
| + /**
|
| + * Specifies number of results to return in a single page. If unspecified, it
|
| + * will default to 256. The maximum value is 2048.
|
| + */
|
| + core.String maxResults;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets. To
|
| + * get the next page of results, set this parameter to the value of
|
| + * "nextPageToken" from the previous response.
|
| + */
|
| + core.String pageToken;
|
| +
|
| + /**
|
| + * The readsets within which to search for reads. At least one readset ID must
|
| + * be provided. All specified readsets must be aligned against a common set of
|
| + * reference sequences; this defines the genomic coordinates for the query.
|
| + */
|
| + core.List<core.String> readsetIds;
|
| +
|
| + /**
|
| + * The end position (1-based, inclusive) of the target range. If specified,
|
| + * "sequenceName" must also be specified. Defaults to the end of the target
|
| + * reference sequence, if any.
|
| + */
|
| + core.String sequenceEnd;
|
| +
|
| + /**
|
| + * Restricts the results to a particular reference sequence such as '1',
|
| + * 'chr1', or 'X'. The set of valid references sequences depends on the
|
| + * readsets specified. If set to "*", only unmapped Reads are returned.
|
| + */
|
| + core.String sequenceName;
|
| +
|
| + /**
|
| + * The start position (1-based, inclusive) of the target range. If specified,
|
| + * "sequenceName" must also be specified. Defaults to the start of the target
|
| + * reference sequence, if any.
|
| + */
|
| + core.String sequenceStart;
|
| +
|
| +
|
| + SearchReadsRequest();
|
| +
|
| + SearchReadsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("maxResults")) {
|
| + maxResults = _json["maxResults"];
|
| + }
|
| + if (_json.containsKey("pageToken")) {
|
| + pageToken = _json["pageToken"];
|
| + }
|
| + if (_json.containsKey("readsetIds")) {
|
| + readsetIds = _json["readsetIds"];
|
| + }
|
| + if (_json.containsKey("sequenceEnd")) {
|
| + sequenceEnd = _json["sequenceEnd"];
|
| + }
|
| + if (_json.containsKey("sequenceName")) {
|
| + sequenceName = _json["sequenceName"];
|
| + }
|
| + if (_json.containsKey("sequenceStart")) {
|
| + sequenceStart = _json["sequenceStart"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (maxResults != null) {
|
| + _json["maxResults"] = maxResults;
|
| + }
|
| + if (pageToken != null) {
|
| + _json["pageToken"] = pageToken;
|
| + }
|
| + if (readsetIds != null) {
|
| + _json["readsetIds"] = readsetIds;
|
| + }
|
| + if (sequenceEnd != null) {
|
| + _json["sequenceEnd"] = sequenceEnd;
|
| + }
|
| + if (sequenceName != null) {
|
| + _json["sequenceName"] = sequenceName;
|
| + }
|
| + if (sequenceStart != null) {
|
| + _json["sequenceStart"] = sequenceStart;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The read search response. */
|
| +class SearchReadsResponse {
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| + /**
|
| + * The list of matching Reads. The resulting Reads are sorted by position; the
|
| + * smallest positions are returned first. Unmapped reads, which have no
|
| + * position, are returned last and are further sorted alphabetically by name.
|
| + */
|
| + core.List<Read> reads;
|
| +
|
| +
|
| + SearchReadsResponse();
|
| +
|
| + SearchReadsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + if (_json.containsKey("reads")) {
|
| + reads = _json["reads"].map((value) => new Read.fromJson(value)).toList();
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + if (reads != null) {
|
| + _json["reads"] = reads.map((value) => (value).toJson()).toList();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset search request. */
|
| +class SearchReadsetsRequest {
|
| + /**
|
| + * Restricts this query to readsets within the given datasets. At least one ID
|
| + * must be provided.
|
| + */
|
| + core.List<core.String> datasetIds;
|
| +
|
| + /**
|
| + * Specifies number of results to return in a single page. If unspecified, it
|
| + * will default to 128. The maximum value is 1024.
|
| + */
|
| + core.String maxResults;
|
| +
|
| + /**
|
| + * Only return readsets for which a substring of the name matches this string.
|
| + */
|
| + core.String name;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets. To
|
| + * get the next page of results, set this parameter to the value of
|
| + * "nextPageToken" from the previous response.
|
| + */
|
| + core.String pageToken;
|
| +
|
| +
|
| + SearchReadsetsRequest();
|
| +
|
| + SearchReadsetsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("datasetIds")) {
|
| + datasetIds = _json["datasetIds"];
|
| + }
|
| + if (_json.containsKey("maxResults")) {
|
| + maxResults = _json["maxResults"];
|
| + }
|
| + if (_json.containsKey("name")) {
|
| + name = _json["name"];
|
| + }
|
| + if (_json.containsKey("pageToken")) {
|
| + pageToken = _json["pageToken"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (datasetIds != null) {
|
| + _json["datasetIds"] = datasetIds;
|
| + }
|
| + if (maxResults != null) {
|
| + _json["maxResults"] = maxResults;
|
| + }
|
| + if (name != null) {
|
| + _json["name"] = name;
|
| + }
|
| + if (pageToken != null) {
|
| + _json["pageToken"] = pageToken;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The readset search response. */
|
| +class SearchReadsetsResponse {
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| + /** The list of matching Readsets. */
|
| + core.List<Readset> readsets;
|
| +
|
| +
|
| + SearchReadsetsResponse();
|
| +
|
| + SearchReadsetsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + if (_json.containsKey("readsets")) {
|
| + readsets = _json["readsets"].map((value) => new Readset.fromJson(value)).toList();
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + if (readsets != null) {
|
| + _json["readsets"] = readsets.map((value) => (value).toJson()).toList();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant search request. */
|
| +class SearchVariantsRequest {
|
| + /**
|
| + * Only return variant calls which belong to callsets with these ids. Leaving
|
| + * this blank returns all variant calls. At most one of callsetNames or
|
| + * callsetIds should be provided.
|
| + */
|
| + core.List<core.String> callsetIds;
|
| +
|
| + /**
|
| + * Only return variant calls which belong to callsets which have exactly these
|
| + * names. Leaving this blank returns all variant calls. At most one of
|
| + * callsetNames or callsetIds should be provided.
|
| + */
|
| + core.List<core.String> callsetNames;
|
| +
|
| + /** Required. Only return variants on this contig. */
|
| + core.String contig;
|
| +
|
| + /**
|
| + * Required. The end of the window (0-based, exclusive) for which overlapping
|
| + * variants should be returned.
|
| + */
|
| + core.String endPosition;
|
| +
|
| + /** The maximum number of variants to return. */
|
| + core.String maxResults;
|
| +
|
| + /**
|
| + * The continuation token, which is used to page through large result sets. To
|
| + * get the next page of results, set this parameter to the value of
|
| + * "nextPageToken" from the previous response.
|
| + */
|
| + core.String pageToken;
|
| +
|
| + /**
|
| + * Required. The beginning of the window (0-based, inclusive) for which
|
| + * overlapping variants should be returned.
|
| + */
|
| + core.String startPosition;
|
| +
|
| + /** Only return variants which have exactly this name. */
|
| + core.String variantName;
|
| +
|
| + /** Required. The ID of the variantset to search. */
|
| + core.String variantsetId;
|
| +
|
| +
|
| + SearchVariantsRequest();
|
| +
|
| + SearchVariantsRequest.fromJson(core.Map _json) {
|
| + if (_json.containsKey("callsetIds")) {
|
| + callsetIds = _json["callsetIds"];
|
| + }
|
| + if (_json.containsKey("callsetNames")) {
|
| + callsetNames = _json["callsetNames"];
|
| + }
|
| + if (_json.containsKey("contig")) {
|
| + contig = _json["contig"];
|
| + }
|
| + if (_json.containsKey("endPosition")) {
|
| + endPosition = _json["endPosition"];
|
| + }
|
| + if (_json.containsKey("maxResults")) {
|
| + maxResults = _json["maxResults"];
|
| + }
|
| + if (_json.containsKey("pageToken")) {
|
| + pageToken = _json["pageToken"];
|
| + }
|
| + if (_json.containsKey("startPosition")) {
|
| + startPosition = _json["startPosition"];
|
| + }
|
| + if (_json.containsKey("variantName")) {
|
| + variantName = _json["variantName"];
|
| + }
|
| + if (_json.containsKey("variantsetId")) {
|
| + variantsetId = _json["variantsetId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (callsetIds != null) {
|
| + _json["callsetIds"] = callsetIds;
|
| + }
|
| + if (callsetNames != null) {
|
| + _json["callsetNames"] = callsetNames;
|
| + }
|
| + if (contig != null) {
|
| + _json["contig"] = contig;
|
| + }
|
| + if (endPosition != null) {
|
| + _json["endPosition"] = endPosition;
|
| + }
|
| + if (maxResults != null) {
|
| + _json["maxResults"] = maxResults;
|
| + }
|
| + if (pageToken != null) {
|
| + _json["pageToken"] = pageToken;
|
| + }
|
| + if (startPosition != null) {
|
| + _json["startPosition"] = startPosition;
|
| + }
|
| + if (variantName != null) {
|
| + _json["variantName"] = variantName;
|
| + }
|
| + if (variantsetId != null) {
|
| + _json["variantsetId"] = variantsetId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/** The variant search response. */
|
| +class SearchVariantsResponse {
|
| + /**
|
| + * The continuation token, which is used to page through large result sets.
|
| + * Provide this value in a subsequent request to return the next page of
|
| + * results. This field will be empty if there aren't any additional results.
|
| + */
|
| + core.String nextPageToken;
|
| +
|
| + /** The list of matching Variants. */
|
| + core.List<Variant> variants;
|
| +
|
| +
|
| + SearchVariantsResponse();
|
| +
|
| + SearchVariantsResponse.fromJson(core.Map _json) {
|
| + if (_json.containsKey("nextPageToken")) {
|
| + nextPageToken = _json["nextPageToken"];
|
| + }
|
| + if (_json.containsKey("variants")) {
|
| + variants = _json["variants"].map((value) => new Variant.fromJson(value)).toList();
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (nextPageToken != null) {
|
| + _json["nextPageToken"] = nextPageToken;
|
| + }
|
| + if (variants != null) {
|
| + _json["variants"] = variants.map((value) => (value).toJson()).toList();
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
| +/**
|
| + * A Variant represents a change in DNA sequence relative to some reference. For
|
| + * example, a Variant could represent a SNP or an insertion. Variants belong to
|
| + * a Variantset.
|
| + */
|
| +class Variant {
|
| + /** The bases that appear instead of the reference bases. */
|
| + core.List<core.String> alternateBases;
|
| +
|
| + /**
|
| + * The variant calls for this particular variant. Each one represents the
|
| + * determination of genotype with respect to this variant.
|
| + */
|
| + core.List<Call> calls;
|
| +
|
| + /** The contig on which this variant occurs. (such as 'chr20' or 'X') */
|
| + core.String contig;
|
| +
|
| + /** The date this variant was created, in milliseconds from the epoch. */
|
| + core.String created;
|
| +
|
| + /**
|
| + * The end position (0-based) of this variant. This corresponds to the first
|
| + * base after the last base in the reference allele. So, the length of the
|
| + * reference allele is (end - position). This is useful for variants that
|
| + * don't explicitly give alternate bases, for example large deletions.
|
| + */
|
| + core.String end;
|
| +
|
| + /** The Google generated ID of the variant, immutable. */
|
| + core.String id;
|
| +
|
| + /** A map of additional variant information. */
|
| + core.Map<core.String, core.List<core.String>> info;
|
| +
|
| + /** Names for the variant, for example a RefSNP ID. */
|
| + core.List<core.String> names;
|
| +
|
| + /**
|
| + * The position at which this variant occurs (0-based). This corresponds to
|
| + * the first base of the string of reference bases.
|
| + */
|
| + core.String position;
|
| +
|
| + /**
|
| + * The reference bases for this variant. They start at the given position.
|
| + */
|
| + core.String referenceBases;
|
| +
|
| + /** The ID of the variantset this variant belongs to. */
|
| + core.String variantsetId;
|
| +
|
| +
|
| + Variant();
|
| +
|
| + Variant.fromJson(core.Map _json) {
|
| + if (_json.containsKey("alternateBases")) {
|
| + alternateBases = _json["alternateBases"];
|
| + }
|
| + if (_json.containsKey("calls")) {
|
| + calls = _json["calls"].map((value) => new Call.fromJson(value)).toList();
|
| + }
|
| + if (_json.containsKey("contig")) {
|
| + contig = _json["contig"];
|
| + }
|
| + if (_json.containsKey("created")) {
|
| + created = _json["created"];
|
| + }
|
| + if (_json.containsKey("end")) {
|
| + end = _json["end"];
|
| + }
|
| + if (_json.containsKey("id")) {
|
| + id = _json["id"];
|
| + }
|
| + if (_json.containsKey("info")) {
|
| + info = _json["info"];
|
| + }
|
| + if (_json.containsKey("names")) {
|
| + names = _json["names"];
|
| + }
|
| + if (_json.containsKey("position")) {
|
| + position = _json["position"];
|
| + }
|
| + if (_json.containsKey("referenceBases")) {
|
| + referenceBases = _json["referenceBases"];
|
| + }
|
| + if (_json.containsKey("variantsetId")) {
|
| + variantsetId = _json["variantsetId"];
|
| + }
|
| + }
|
| +
|
| + core.Map toJson() {
|
| + var _json = new core.Map();
|
| + if (alternateBases != null) {
|
| + _json["alternateBases"] = alternateBases;
|
| + }
|
| + if (calls != null) {
|
| + _json["calls"] = calls.map((value) => (value).toJson()).toList();
|
| + }
|
| + if (contig != null) {
|
| + _json["contig"] = contig;
|
| + }
|
| + if (created != null) {
|
| + _json["created"] = created;
|
| + }
|
| + if (end != null) {
|
| + _json["end"] = end;
|
| + }
|
| + if (id != null) {
|
| + _json["id"] = id;
|
| + }
|
| + if (info != null) {
|
| + _json["info"] = info;
|
| + }
|
| + if (names != null) {
|
| + _json["names"] = names;
|
| + }
|
| + if (position != null) {
|
| + _json["position"] = position;
|
| + }
|
| + if (referenceBases != null) {
|
| + _json["referenceBases"] = referenceBases;
|
| + }
|
| + if (variantsetId != null) {
|
| + _json["variantsetId"] = variantsetId;
|
| + }
|
| + return _json;
|
| + }
|
| +}
|
| +
|
| +
|
|
|