Index: generated/googleapis_beta/lib/genomics/v1beta2.dart |
diff --git a/generated/googleapis_beta/lib/genomics/v1beta.dart b/generated/googleapis_beta/lib/genomics/v1beta2.dart |
similarity index 61% |
copy from generated/googleapis_beta/lib/genomics/v1beta.dart |
copy to generated/googleapis_beta/lib/genomics/v1beta2.dart |
index ef44652c806d243bdd7430cf2d0a99b5c056124b..a4ab8bbfed2451fb299089c15b4b97bc8fa9420a 100644 |
--- a/generated/googleapis_beta/lib/genomics/v1beta.dart |
+++ b/generated/googleapis_beta/lib/genomics/v1beta2.dart |
@@ -1,4 +1,4 @@ |
-library googleapis_beta.genomics.v1beta; |
+library googleapis_beta.genomics.v1beta2; |
import "dart:core" as core; |
import "dart:collection" as collection; |
@@ -30,85 +30,19 @@ class GenomicsApi { |
final common_internal.ApiRequester _requester; |
- BeaconsResourceApi get beacons => new BeaconsResourceApi(_requester); |
CallsetsResourceApi get callsets => new CallsetsResourceApi(_requester); |
DatasetsResourceApi get datasets => new DatasetsResourceApi(_requester); |
ExperimentalResourceApi get experimental => new ExperimentalResourceApi(_requester); |
JobsResourceApi get jobs => new JobsResourceApi(_requester); |
+ ReadgroupsetsResourceApi get readgroupsets => new ReadgroupsetsResourceApi(_requester); |
ReadsResourceApi get reads => new ReadsResourceApi(_requester); |
- ReadsetsResourceApi get readsets => new ReadsetsResourceApi(_requester); |
+ ReferencesResourceApi get references => new ReferencesResourceApi(_requester); |
+ ReferencesetsResourceApi get referencesets => new ReferencesetsResourceApi(_requester); |
VariantsResourceApi get variants => new VariantsResourceApi(_requester); |
VariantsetsResourceApi get variantsets => new VariantsetsResourceApi(_requester); |
GenomicsApi(http.Client client) : |
- _requester = new common_internal.ApiRequester(client, "https://www.googleapis.com/", "genomics/v1beta/"); |
-} |
- |
- |
-/** Not documented yet. */ |
-class BeaconsResourceApi { |
- final common_internal.ApiRequester _requester; |
- |
- BeaconsResourceApi(common_internal.ApiRequester client) : |
- _requester = client; |
- |
- /** |
- * This is an experimental API that provides a Global Alliance for Genomics |
- * and Health Beacon. It may change at any time. |
- * |
- * Request parameters: |
- * |
- * [variantSetId] - The ID of the variant set to query over. It must be |
- * public. Private variant sets will return an unauthorized exception. |
- * |
- * [allele] - Required. The allele to look for ('A', 'C', 'G' or 'T'). |
- * |
- * [position] - Required. The 0-based position to query. |
- * |
- * [referenceName] - Required. The reference to query over. |
- * |
- * Completes with a [Beacon]. |
- * |
- * Completes with a [common.ApiRequestError] if the API endpoint returned an |
- * error. |
- * |
- * If the used [http.Client] completes with an error when making a REST call, |
- * this method will complete with the same error. |
- */ |
- async.Future<Beacon> get(core.String variantSetId, {core.String allele, core.String position, core.String referenceName}) { |
- var _url = null; |
- var _queryParams = new core.Map(); |
- var _uploadMedia = null; |
- var _uploadOptions = null; |
- var _downloadOptions = common.DownloadOptions.Metadata; |
- var _body = null; |
- |
- if (variantSetId == null) { |
- throw new core.ArgumentError("Parameter variantSetId is required."); |
- } |
- if (allele != null) { |
- _queryParams["allele"] = [allele]; |
- } |
- if (position != null) { |
- _queryParams["position"] = [position]; |
- } |
- if (referenceName != null) { |
- _queryParams["referenceName"] = [referenceName]; |
- } |
- |
- |
- _url = 'beacons/' + common_internal.Escaper.ecapeVariable('$variantSetId'); |
- |
- var _response = _requester.request(_url, |
- "GET", |
- body: _body, |
- queryParams: _queryParams, |
- uploadOptions: _uploadOptions, |
- uploadMedia: _uploadMedia, |
- downloadOptions: _downloadOptions); |
- return _response.then((data) => new Beacon.fromJson(data)); |
- } |
- |
+ _requester = new common_internal.ApiRequester(client, "https://www.googleapis.com/", "genomics/v1beta2/"); |
} |
@@ -164,7 +98,7 @@ class CallsetsResourceApi { |
* |
* Request parameters: |
* |
- * [callSetId] - The ID of the callset to be deleted. |
+ * [callSetId] - The ID of the call set to be deleted. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -203,7 +137,7 @@ class CallsetsResourceApi { |
* |
* Request parameters: |
* |
- * [callSetId] - The ID of the callset. |
+ * [callSetId] - The ID of the call set. |
* |
* Completes with a [CallSet]. |
* |
@@ -245,7 +179,7 @@ class CallsetsResourceApi { |
* |
* Request parameters: |
* |
- * [callSetId] - The ID of the callset to be updated. |
+ * [callSetId] - The ID of the call set to be updated. |
* |
* Completes with a [CallSet]. |
* |
@@ -286,6 +220,8 @@ class CallsetsResourceApi { |
/** |
* Gets a list of call sets matching the criteria. |
* |
+ * Implements GlobalAllianceApi.searchCallSets. |
+ * |
* [request] - The metadata request object. |
* |
* Request parameters: |
@@ -330,7 +266,7 @@ class CallsetsResourceApi { |
* |
* Request parameters: |
* |
- * [callSetId] - The ID of the callset to be updated. |
+ * [callSetId] - The ID of the call set to be updated. |
* |
* Completes with a [CallSet]. |
* |
@@ -502,15 +438,15 @@ class DatasetsResourceApi { |
* |
* Request parameters: |
* |
- * [maxResults] - The maximum number of results returned by this request. |
+ * [pageSize] - The maximum number of results returned by this request. |
* |
* [pageToken] - The continuation token, which is used to page through large |
* result sets. To get the next page of results, set this parameter to the |
* value of nextPageToken from the previous response. |
* |
- * [projectId] - Only return datasets which belong to this Google Developers |
- * Console project. Only accepts project numbers. Returns all public projects |
- * if no project number is specified. |
+ * [projectNumber] - Only return datasets which belong to this Google |
+ * Developers Console project. Only accepts project numbers. Returns all |
+ * public projects if no project number is specified. |
* |
* Completes with a [ListDatasetsResponse]. |
* |
@@ -520,7 +456,7 @@ class DatasetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ListDatasetsResponse> list({core.String maxResults, core.String pageToken, core.String projectId}) { |
+ async.Future<ListDatasetsResponse> list({core.int pageSize, core.String pageToken, core.String projectNumber}) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -528,14 +464,14 @@ class DatasetsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (maxResults != null) { |
- _queryParams["maxResults"] = [maxResults]; |
+ if (pageSize != null) { |
+ _queryParams["pageSize"] = ["${pageSize}"]; |
} |
if (pageToken != null) { |
_queryParams["pageToken"] = [pageToken]; |
} |
- if (projectId != null) { |
- _queryParams["projectId"] = [projectId]; |
+ if (projectNumber != null) { |
+ _queryParams["projectNumber"] = [projectNumber]; |
} |
@@ -878,28 +814,24 @@ class JobsResourceApi { |
/** Not documented yet. */ |
-class ReadsResourceApi { |
+class ReadgroupsetsResourceApi { |
final common_internal.ApiRequester _requester; |
- ReadsResourceApi(common_internal.ApiRequester client) : |
+ ReadgroupsetsCoveragebucketsResourceApi get coveragebuckets => new ReadgroupsetsCoveragebucketsResourceApi(_requester); |
+ |
+ ReadgroupsetsResourceApi(common_internal.ApiRequester client) : |
_requester = client; |
/** |
- * Gets a list of reads for one or more readsets. Reads search operates over a |
- * genomic coordinate space of reference sequence & position defined over the |
- * reference sequences to which the requested readsets are aligned. If a |
- * target positional range is specified, search returns all reads whose |
- * alignment to the reference genome overlap the range. A query which |
- * specifies only readset IDs yields all reads in those readsets, including |
- * unmapped reads. All reads returned (including reads on subsequent pages) |
- * are ordered by genomic coordinate (reference sequence & position). Reads |
- * with equivalent genomic coordinates are returned in a deterministic order. |
+ * Aligns read data from existing read group sets or files from Google Cloud |
+ * Storage. See the alignment and variant calling documentation for more |
+ * details. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [SearchReadsResponse]. |
+ * Completes with a [AlignReadGroupSetsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -907,7 +839,7 @@ class ReadsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<SearchReadsResponse> search(SearchReadsRequest request) { |
+ async.Future<AlignReadGroupSetsResponse> align(AlignReadGroupSetsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -920,7 +852,7 @@ class ReadsResourceApi { |
} |
- _url = 'reads/search'; |
+ _url = 'readgroupsets/align'; |
var _response = _requester.request(_url, |
"POST", |
@@ -929,28 +861,59 @@ class ReadsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new SearchReadsResponse.fromJson(data)); |
+ return _response.then((data) => new AlignReadGroupSetsResponse.fromJson(data)); |
} |
-} |
+ /** |
+ * Calls variants on read data from existing read group sets or files from |
+ * Google Cloud Storage. See the alignment and variant calling documentation |
+ * for more details. |
+ * |
+ * [request] - The metadata request object. |
+ * |
+ * Request parameters: |
+ * |
+ * Completes with a [CallReadGroupSetsResponse]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<CallReadGroupSetsResponse> call(CallReadGroupSetsRequest request) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ if (request != null) { |
+ _body = convert.JSON.encode((request).toJson()); |
+ } |
-/** Not documented yet. */ |
-class ReadsetsResourceApi { |
- final common_internal.ApiRequester _requester; |
- ReadsetsCoveragebucketsResourceApi get coveragebuckets => new ReadsetsCoveragebucketsResourceApi(_requester); |
+ _url = 'readgroupsets/call'; |
- ReadsetsResourceApi(common_internal.ApiRequester client) : |
- _requester = client; |
+ var _response = _requester.request(_url, |
+ "POST", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new CallReadGroupSetsResponse.fromJson(data)); |
+ } |
/** |
- * Deletes a readset. |
+ * Deletes a read group set. |
* |
* Request parameters: |
* |
- * [readsetId] - The ID of the readset to be deleted. The caller must have |
- * WRITE permissions to the dataset associated with this readset. |
+ * [readGroupSetId] - The ID of the read group set to be deleted. The caller |
+ * must have WRITE permissions to the dataset associated with this read group |
+ * set. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -958,7 +921,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future delete(core.String readsetId) { |
+ async.Future delete(core.String readGroupSetId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -966,13 +929,13 @@ class ReadsetsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (readsetId == null) { |
- throw new core.ArgumentError("Parameter readsetId is required."); |
+ if (readGroupSetId == null) { |
+ throw new core.ArgumentError("Parameter readGroupSetId is required."); |
} |
_downloadOptions = null; |
- _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId'); |
+ _url = 'readgroupsets/' + common_internal.Escaper.ecapeVariable('$readGroupSetId'); |
var _response = _requester.request(_url, |
"DELETE", |
@@ -985,17 +948,18 @@ class ReadsetsResourceApi { |
} |
/** |
- * Exports readsets to a BAM file in Google Cloud Storage. Note that currently |
- * there may be some differences between exported BAM files and the original |
- * BAM file at the time of import. In particular, comments in the input file |
- * header will not be preserved, and some custom tags will be converted to |
- * strings. |
+ * Exports read group sets to a BAM file in Google Cloud Storage. |
+ * |
+ * Note that currently there may be some differences between exported BAM |
+ * files and the original BAM file at the time of import. In particular, |
+ * comments in the input file header will not be preserved, and some custom |
+ * tags will be converted to strings. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [ExportReadsetsResponse]. |
+ * Completes with a [ExportReadGroupSetsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1003,7 +967,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ExportReadsetsResponse> export(ExportReadsetsRequest request) { |
+ async.Future<ExportReadGroupSetsResponse> export(ExportReadGroupSetsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1016,7 +980,7 @@ class ReadsetsResourceApi { |
} |
- _url = 'readsets/export'; |
+ _url = 'readgroupsets/export'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1025,17 +989,17 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new ExportReadsetsResponse.fromJson(data)); |
+ return _response.then((data) => new ExportReadGroupSetsResponse.fromJson(data)); |
} |
/** |
- * Gets a readset by ID. |
+ * Gets a read group set by ID. |
* |
* Request parameters: |
* |
- * [readsetId] - The ID of the readset. |
+ * [readGroupSetId] - The ID of the read group set. |
* |
- * Completes with a [Readset]. |
+ * Completes with a [ReadGroupSet]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1043,7 +1007,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Readset> get(core.String readsetId) { |
+ async.Future<ReadGroupSet> get(core.String readGroupSetId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1051,12 +1015,12 @@ class ReadsetsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (readsetId == null) { |
- throw new core.ArgumentError("Parameter readsetId is required."); |
+ if (readGroupSetId == null) { |
+ throw new core.ArgumentError("Parameter readGroupSetId is required."); |
} |
- _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId'); |
+ _url = 'readgroupsets/' + common_internal.Escaper.ecapeVariable('$readGroupSetId'); |
var _response = _requester.request(_url, |
"GET", |
@@ -1065,20 +1029,22 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new Readset.fromJson(data)); |
+ return _response.then((data) => new ReadGroupSet.fromJson(data)); |
} |
/** |
- * Creates readsets by asynchronously importing the provided information. Note |
- * that currently comments in the input file header are not imported and some |
- * custom tags will be converted to strings, rather than preserving tag types. |
- * The caller must have WRITE permissions to the dataset. |
+ * Creates read group sets by asynchronously importing the provided |
+ * information. |
+ * |
+ * Note that currently comments in the input file header are not imported and |
+ * some custom tags will be converted to strings, rather than preserving tag |
+ * types. The caller must have WRITE permissions to the dataset. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [ImportReadsetsResponse]. |
+ * Completes with a [ImportReadGroupSetsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1086,7 +1052,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ImportReadsetsResponse> import(ImportReadsetsRequest request) { |
+ async.Future<ImportReadGroupSetsResponse> import(ImportReadGroupSetsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1099,7 +1065,7 @@ class ReadsetsResourceApi { |
} |
- _url = 'readsets/import'; |
+ _url = 'readgroupsets/import'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1108,20 +1074,21 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new ImportReadsetsResponse.fromJson(data)); |
+ return _response.then((data) => new ImportReadGroupSetsResponse.fromJson(data)); |
} |
/** |
- * Updates a readset. This method supports patch semantics. |
+ * Updates a read group set. This method supports patch semantics. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * [readsetId] - The ID of the readset to be updated. The caller must have |
- * WRITE permissions to the dataset associated with this readset. |
+ * [readGroupSetId] - The ID of the read group set to be updated. The caller |
+ * must have WRITE permissions to the dataset associated with this read group |
+ * set. |
* |
- * Completes with a [Readset]. |
+ * Completes with a [ReadGroupSet]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1129,7 +1096,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Readset> patch(Readset request, core.String readsetId) { |
+ async.Future<ReadGroupSet> patch(ReadGroupSet request, core.String readGroupSetId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1140,12 +1107,12 @@ class ReadsetsResourceApi { |
if (request != null) { |
_body = convert.JSON.encode((request).toJson()); |
} |
- if (readsetId == null) { |
- throw new core.ArgumentError("Parameter readsetId is required."); |
+ if (readGroupSetId == null) { |
+ throw new core.ArgumentError("Parameter readGroupSetId is required."); |
} |
- _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId'); |
+ _url = 'readgroupsets/' + common_internal.Escaper.ecapeVariable('$readGroupSetId'); |
var _response = _requester.request(_url, |
"PATCH", |
@@ -1154,17 +1121,19 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new Readset.fromJson(data)); |
+ return _response.then((data) => new ReadGroupSet.fromJson(data)); |
} |
/** |
- * Gets a list of readsets matching the criteria. |
+ * Searches for read group sets matching the criteria. |
+ * |
+ * Implements GlobalAllianceApi.searchReadGroupSets. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [SearchReadsetsResponse]. |
+ * Completes with a [SearchReadGroupSetsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1172,7 +1141,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<SearchReadsetsResponse> search(SearchReadsetsRequest request) { |
+ async.Future<SearchReadGroupSetsResponse> search(SearchReadGroupSetsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1185,7 +1154,7 @@ class ReadsetsResourceApi { |
} |
- _url = 'readsets/search'; |
+ _url = 'readgroupsets/search'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1194,20 +1163,21 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new SearchReadsetsResponse.fromJson(data)); |
+ return _response.then((data) => new SearchReadGroupSetsResponse.fromJson(data)); |
} |
/** |
- * Updates a readset. |
+ * Updates a read group set. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * [readsetId] - The ID of the readset to be updated. The caller must have |
- * WRITE permissions to the dataset associated with this readset. |
+ * [readGroupSetId] - The ID of the read group set to be updated. The caller |
+ * must have WRITE permissions to the dataset associated with this read group |
+ * set. |
* |
- * Completes with a [Readset]. |
+ * Completes with a [ReadGroupSet]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1215,7 +1185,7 @@ class ReadsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Readset> update(Readset request, core.String readsetId) { |
+ async.Future<ReadGroupSet> update(ReadGroupSet request, core.String readGroupSetId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1226,12 +1196,12 @@ class ReadsetsResourceApi { |
if (request != null) { |
_body = convert.JSON.encode((request).toJson()); |
} |
- if (readsetId == null) { |
- throw new core.ArgumentError("Parameter readsetId is required."); |
+ if (readGroupSetId == null) { |
+ throw new core.ArgumentError("Parameter readGroupSetId is required."); |
} |
- _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId'); |
+ _url = 'readgroupsets/' + common_internal.Escaper.ecapeVariable('$readGroupSetId'); |
var _response = _requester.request(_url, |
"PUT", |
@@ -1240,48 +1210,50 @@ class ReadsetsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new Readset.fromJson(data)); |
+ return _response.then((data) => new ReadGroupSet.fromJson(data)); |
} |
} |
/** Not documented yet. */ |
-class ReadsetsCoveragebucketsResourceApi { |
+class ReadgroupsetsCoveragebucketsResourceApi { |
final common_internal.ApiRequester _requester; |
- ReadsetsCoveragebucketsResourceApi(common_internal.ApiRequester client) : |
+ ReadgroupsetsCoveragebucketsResourceApi(common_internal.ApiRequester client) : |
_requester = client; |
/** |
- * Lists fixed width coverage buckets for a readset, each of which correspond |
- * to a range of a reference sequence. Each bucket summarizes coverage |
- * information across its corresponding genomic range. Coverage is defined as |
- * the number of reads which are aligned to a given base in the reference |
- * sequence. Coverage buckets are available at various bucket widths, enabling |
- * various coverage "zoom levels". The caller must have READ permissions for |
- * the target readset. |
+ * Lists fixed width coverage buckets for a read group set, each of which |
+ * correspond to a range of a reference sequence. Each bucket summarizes |
+ * coverage information across its corresponding genomic range. |
+ * |
+ * Coverage is defined as the number of reads which are aligned to a given |
+ * base in the reference sequence. Coverage buckets are available at several |
+ * precomputed bucket widths, enabling retrieval of various coverage 'zoom |
+ * levels'. The caller must have READ permissions for the target read group |
+ * set. |
* |
* Request parameters: |
* |
- * [readsetId] - Required. The ID of the readset over which coverage is |
- * requested. |
+ * [readGroupSetId] - Required. The ID of the read group set over which |
+ * coverage is requested. |
* |
- * [maxResults] - The maximum number of results to return in a single page. If |
+ * [pageSize] - The maximum number of results to return in a single page. If |
* unspecified, defaults to 1024. The maximum value is 2048. |
* |
* [pageToken] - The continuation token, which is used to page through large |
* result sets. To get the next page of results, set this parameter to the |
* value of nextPageToken from the previous response. |
* |
- * [range_sequenceEnd] - The end position of the range on the reference, |
- * 1-based exclusive. If specified, sequenceName must also be specified. |
+ * [range_end] - The end position of the range on the reference, 0-based |
+ * exclusive. If specified, referenceName must also be specified. |
* |
- * [range_sequenceName] - The reference sequence name, for example chr1, 1, or |
- * chrX. |
+ * [range_referenceName] - The reference sequence name, for example chr1, 1, |
+ * or chrX. |
* |
- * [range_sequenceStart] - The start position of the range on the reference, |
- * 1-based inclusive. If specified, sequenceName must also be specified. |
+ * [range_start] - The start position of the range on the reference, 0-based |
+ * inclusive. If specified, referenceName must also be specified. |
* |
* [targetBucketWidth] - The desired width of each reported coverage bucket in |
* base pairs. This will be rounded down to the nearest precomputed bucket |
@@ -1298,7 +1270,7 @@ class ReadsetsCoveragebucketsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ListCoverageBucketsResponse> list(core.String readsetId, {core.String maxResults, core.String pageToken, core.String range_sequenceEnd, core.String range_sequenceName, core.String range_sequenceStart, core.String targetBucketWidth}) { |
+ async.Future<ListCoverageBucketsResponse> list(core.String readGroupSetId, {core.int pageSize, core.String pageToken, core.String range_end, core.String range_referenceName, core.String range_start, core.String targetBucketWidth}) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1306,30 +1278,30 @@ class ReadsetsCoveragebucketsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (readsetId == null) { |
- throw new core.ArgumentError("Parameter readsetId is required."); |
+ if (readGroupSetId == null) { |
+ throw new core.ArgumentError("Parameter readGroupSetId is required."); |
} |
- if (maxResults != null) { |
- _queryParams["maxResults"] = [maxResults]; |
+ if (pageSize != null) { |
+ _queryParams["pageSize"] = ["${pageSize}"]; |
} |
if (pageToken != null) { |
_queryParams["pageToken"] = [pageToken]; |
} |
- if (range_sequenceEnd != null) { |
- _queryParams["range.sequenceEnd"] = [range_sequenceEnd]; |
+ if (range_end != null) { |
+ _queryParams["range.end"] = [range_end]; |
} |
- if (range_sequenceName != null) { |
- _queryParams["range.sequenceName"] = [range_sequenceName]; |
+ if (range_referenceName != null) { |
+ _queryParams["range.referenceName"] = [range_referenceName]; |
} |
- if (range_sequenceStart != null) { |
- _queryParams["range.sequenceStart"] = [range_sequenceStart]; |
+ if (range_start != null) { |
+ _queryParams["range.start"] = [range_start]; |
} |
if (targetBucketWidth != null) { |
_queryParams["targetBucketWidth"] = [targetBucketWidth]; |
} |
- _url = 'readsets/' + common_internal.Escaper.ecapeVariable('$readsetId') + '/coveragebuckets'; |
+ _url = 'readgroupsets/' + common_internal.Escaper.ecapeVariable('$readGroupSetId') + '/coveragebuckets'; |
var _response = _requester.request(_url, |
"GET", |
@@ -1345,20 +1317,34 @@ class ReadsetsCoveragebucketsResourceApi { |
/** Not documented yet. */ |
-class VariantsResourceApi { |
+class ReadsResourceApi { |
final common_internal.ApiRequester _requester; |
- VariantsResourceApi(common_internal.ApiRequester client) : |
+ ReadsResourceApi(common_internal.ApiRequester client) : |
_requester = client; |
/** |
- * Creates a new variant. |
+ * Gets a list of reads for one or more read group sets. Reads search operates |
+ * over a genomic coordinate space of reference sequence & position defined |
+ * over the reference sequences to which the requested read group sets are |
+ * aligned. |
+ * |
+ * If a target positional range is specified, search returns all reads whose |
+ * alignment to the reference genome overlap the range. A query which |
+ * specifies only read group set IDs yields all reads in those read group |
+ * sets, including unmapped reads. |
+ * |
+ * All reads returned (including reads on subsequent pages) are ordered by |
+ * genomic coordinate (reference sequence & position). Reads with equivalent |
+ * genomic coordinates are returned in a deterministic order. |
+ * |
+ * Implements GlobalAllianceApi.searchReads. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [Variant]. |
+ * Completes with a [SearchReadsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1366,7 +1352,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Variant> create(Variant request) { |
+ async.Future<SearchReadsResponse> search(SearchReadsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1379,7 +1365,7 @@ class VariantsResourceApi { |
} |
- _url = 'variants'; |
+ _url = 'reads/search'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1388,15 +1374,31 @@ class VariantsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new Variant.fromJson(data)); |
+ return _response.then((data) => new SearchReadsResponse.fromJson(data)); |
} |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ReferencesResourceApi { |
+ final common_internal.ApiRequester _requester; |
+ |
+ ReferencesBasesResourceApi get bases => new ReferencesBasesResourceApi(_requester); |
+ |
+ ReferencesResourceApi(common_internal.ApiRequester client) : |
+ _requester = client; |
+ |
/** |
- * Deletes a variant. |
+ * Gets a reference. |
+ * |
+ * Implements GlobalAllianceApi.getReference. |
* |
* Request parameters: |
* |
- * [variantId] - The ID of the variant to be deleted. |
+ * [referenceId] - The ID of the reference. |
+ * |
+ * Completes with a [Reference]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1404,7 +1406,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future delete(core.String variantId) { |
+ async.Future<Reference> get(core.String referenceId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1412,32 +1414,33 @@ class VariantsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (variantId == null) { |
- throw new core.ArgumentError("Parameter variantId is required."); |
+ if (referenceId == null) { |
+ throw new core.ArgumentError("Parameter referenceId is required."); |
} |
- _downloadOptions = null; |
- _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
+ _url = 'references/' + common_internal.Escaper.ecapeVariable('$referenceId'); |
var _response = _requester.request(_url, |
- "DELETE", |
+ "GET", |
body: _body, |
queryParams: _queryParams, |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => null); |
+ return _response.then((data) => new Reference.fromJson(data)); |
} |
/** |
- * Exports variant data to an external destination. |
+ * Searches for references which match the given criteria. |
+ * |
+ * Implements GlobalAllianceApi.searchReferences. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [ExportVariantsResponse]. |
+ * Completes with a [SearchReferencesResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1445,7 +1448,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ExportVariantsResponse> export(ExportVariantsRequest request) { |
+ async.Future<SearchReferencesResponse> search(SearchReferencesRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1458,7 +1461,7 @@ class VariantsResourceApi { |
} |
- _url = 'variants/export'; |
+ _url = 'references/search'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1467,17 +1470,41 @@ class VariantsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new ExportVariantsResponse.fromJson(data)); |
+ return _response.then((data) => new SearchReferencesResponse.fromJson(data)); |
} |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ReferencesBasesResourceApi { |
+ final common_internal.ApiRequester _requester; |
+ |
+ ReferencesBasesResourceApi(common_internal.ApiRequester client) : |
+ _requester = client; |
+ |
/** |
- * Gets a variant by ID. |
+ * Lists the bases in a reference, optionally restricted to a range. |
+ * |
+ * Implements GlobalAllianceApi.getReferenceBases. |
* |
* Request parameters: |
* |
- * [variantId] - The ID of the variant. |
+ * [referenceId] - The ID of the reference. |
* |
- * Completes with a [Variant]. |
+ * [end] - The end position (0-based, exclusive) of this query. Defaults to |
+ * the length of this reference. |
+ * |
+ * [pageSize] - Specifies the maximum number of bases to return in a single |
+ * page. |
+ * |
+ * [pageToken] - The continuation token, which is used to page through large |
+ * result sets. To get the next page of results, set this parameter to the |
+ * value of nextPageToken from the previous response. |
+ * |
+ * [start] - The start position (0-based) of this query. Defaults to 0. |
+ * |
+ * Completes with a [ListBasesResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1485,7 +1512,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Variant> get(core.String variantId) { |
+ async.Future<ListBasesResponse> list(core.String referenceId, {core.String end, core.int pageSize, core.String pageToken, core.String start}) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1493,12 +1520,24 @@ class VariantsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (variantId == null) { |
- throw new core.ArgumentError("Parameter variantId is required."); |
+ if (referenceId == null) { |
+ throw new core.ArgumentError("Parameter referenceId is required."); |
+ } |
+ if (end != null) { |
+ _queryParams["end"] = [end]; |
+ } |
+ if (pageSize != null) { |
+ _queryParams["pageSize"] = ["${pageSize}"]; |
+ } |
+ if (pageToken != null) { |
+ _queryParams["pageToken"] = [pageToken]; |
+ } |
+ if (start != null) { |
+ _queryParams["start"] = [start]; |
} |
- _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
+ _url = 'references/' + common_internal.Escaper.ecapeVariable('$referenceId') + '/bases'; |
var _response = _requester.request(_url, |
"GET", |
@@ -1507,24 +1546,29 @@ class VariantsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new Variant.fromJson(data)); |
+ return _response.then((data) => new ListBasesResponse.fromJson(data)); |
} |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ReferencesetsResourceApi { |
+ final common_internal.ApiRequester _requester; |
+ |
+ ReferencesetsResourceApi(common_internal.ApiRequester client) : |
+ _requester = client; |
+ |
/** |
- * Creates variant data by asynchronously importing the provided information. |
- * The variants for import will be merged with any existing data and each |
- * other according to the behavior of mergeVariants. In particular, this means |
- * for merged VCF variants that have conflicting INFO fields, some data will |
- * be arbitrarily discarded. As a special case, for single-sample VCF files, |
- * QUAL and FILTER fields will be moved to the call level; these are sometimes |
- * interpreted in a call-specific context. Imported VCF headers are appended |
- * to the metadata already in a VariantSet. |
+ * Gets a reference set. |
* |
- * [request] - The metadata request object. |
+ * Implements GlobalAllianceApi.getReferenceSet. |
* |
* Request parameters: |
* |
- * Completes with a [ImportVariantsResponse]. |
+ * [referenceSetId] - The ID of the reference set. |
+ * |
+ * Completes with a [ReferenceSet]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1532,7 +1576,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<ImportVariantsResponse> import(ImportVariantsRequest request) { |
+ async.Future<ReferenceSet> get(core.String referenceSetId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1540,31 +1584,33 @@ class VariantsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (request != null) { |
- _body = convert.JSON.encode((request).toJson()); |
+ if (referenceSetId == null) { |
+ throw new core.ArgumentError("Parameter referenceSetId is required."); |
} |
- _url = 'variants/import'; |
+ _url = 'referencesets/' + common_internal.Escaper.ecapeVariable('$referenceSetId'); |
var _response = _requester.request(_url, |
- "POST", |
+ "GET", |
body: _body, |
queryParams: _queryParams, |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new ImportVariantsResponse.fromJson(data)); |
+ return _response.then((data) => new ReferenceSet.fromJson(data)); |
} |
/** |
- * Gets a list of variants matching the criteria. |
+ * Searches for reference sets which match the given criteria. |
+ * |
+ * Implements GlobalAllianceApi.searchReferenceSets. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * Completes with a [SearchVariantsResponse]. |
+ * Completes with a [SearchReferenceSetsResponse]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1572,7 +1618,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<SearchVariantsResponse> search(SearchVariantsRequest request) { |
+ async.Future<SearchReferenceSetsResponse> search(SearchReferenceSetsRequest request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1585,7 +1631,7 @@ class VariantsResourceApi { |
} |
- _url = 'variants/search'; |
+ _url = 'referencesets/search'; |
var _response = _requester.request(_url, |
"POST", |
@@ -1594,19 +1640,26 @@ class VariantsResourceApi { |
uploadOptions: _uploadOptions, |
uploadMedia: _uploadMedia, |
downloadOptions: _downloadOptions); |
- return _response.then((data) => new SearchVariantsResponse.fromJson(data)); |
+ return _response.then((data) => new SearchReferenceSetsResponse.fromJson(data)); |
} |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class VariantsResourceApi { |
+ final common_internal.ApiRequester _requester; |
+ |
+ VariantsResourceApi(common_internal.ApiRequester client) : |
+ _requester = client; |
+ |
/** |
- * Updates a variant's names and info fields. All other modifications are |
- * silently ignored. Returns the modified variant without its calls. |
+ * Creates a new variant. |
* |
* [request] - The metadata request object. |
* |
* Request parameters: |
* |
- * [variantId] - The ID of the variant to be updated. |
- * |
* Completes with a [Variant]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
@@ -1615,7 +1668,7 @@ class VariantsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<Variant> update(Variant request, core.String variantId) { |
+ async.Future<Variant> create(Variant request) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1626,15 +1679,12 @@ class VariantsResourceApi { |
if (request != null) { |
_body = convert.JSON.encode((request).toJson()); |
} |
- if (variantId == null) { |
- throw new core.ArgumentError("Parameter variantId is required."); |
- } |
- _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
+ _url = 'variants'; |
var _response = _requester.request(_url, |
- "PUT", |
+ "POST", |
body: _body, |
queryParams: _queryParams, |
uploadOptions: _uploadOptions, |
@@ -1643,23 +1693,12 @@ class VariantsResourceApi { |
return _response.then((data) => new Variant.fromJson(data)); |
} |
-} |
- |
- |
-/** Not documented yet. */ |
-class VariantsetsResourceApi { |
- final common_internal.ApiRequester _requester; |
- |
- VariantsetsResourceApi(common_internal.ApiRequester client) : |
- _requester = client; |
- |
/** |
- * Deletes the contents of a variant set. The variant set object is not |
- * deleted. |
+ * Deletes a variant. |
* |
* Request parameters: |
* |
- * [variantSetId] - The ID of the variant set to be deleted. |
+ * [variantId] - The ID of the variant to be deleted. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1667,7 +1706,7 @@ class VariantsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future delete(core.String variantSetId) { |
+ async.Future delete(core.String variantId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1675,13 +1714,13 @@ class VariantsetsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (variantSetId == null) { |
- throw new core.ArgumentError("Parameter variantSetId is required."); |
+ if (variantId == null) { |
+ throw new core.ArgumentError("Parameter variantId is required."); |
} |
_downloadOptions = null; |
- _url = 'variantsets/' + common_internal.Escaper.ecapeVariable('$variantSetId'); |
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
var _response = _requester.request(_url, |
"DELETE", |
@@ -1694,13 +1733,13 @@ class VariantsetsResourceApi { |
} |
/** |
- * Gets a variant set by ID. |
+ * Gets a variant by ID. |
* |
* Request parameters: |
* |
- * [variantSetId] - Required. The ID of the variant set. |
+ * [variantId] - The ID of the variant. |
* |
- * Completes with a [VariantSet]. |
+ * Completes with a [Variant]. |
* |
* Completes with a [common.ApiRequestError] if the API endpoint returned an |
* error. |
@@ -1708,7 +1747,7 @@ class VariantsetsResourceApi { |
* If the used [http.Client] completes with an error when making a REST call, |
* this method will complete with the same error. |
*/ |
- async.Future<VariantSet> get(core.String variantSetId) { |
+ async.Future<Variant> get(core.String variantId) { |
var _url = null; |
var _queryParams = new core.Map(); |
var _uploadMedia = null; |
@@ -1716,7 +1755,232 @@ class VariantsetsResourceApi { |
var _downloadOptions = common.DownloadOptions.Metadata; |
var _body = null; |
- if (variantSetId == null) { |
+ if (variantId == null) { |
+ throw new core.ArgumentError("Parameter variantId is required."); |
+ } |
+ |
+ |
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
+ |
+ var _response = _requester.request(_url, |
+ "GET", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new Variant.fromJson(data)); |
+ } |
+ |
+ /** |
+ * Gets a list of variants matching the criteria. |
+ * |
+ * Implements GlobalAllianceApi.searchVariants. |
+ * |
+ * [request] - The metadata request object. |
+ * |
+ * Request parameters: |
+ * |
+ * Completes with a [SearchVariantsResponse]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<SearchVariantsResponse> search(SearchVariantsRequest request) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (request != null) { |
+ _body = convert.JSON.encode((request).toJson()); |
+ } |
+ |
+ |
+ _url = 'variants/search'; |
+ |
+ var _response = _requester.request(_url, |
+ "POST", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new SearchVariantsResponse.fromJson(data)); |
+ } |
+ |
+ /** |
+ * Updates a variant's names and info fields. All other modifications are |
+ * silently ignored. Returns the modified variant without its calls. |
+ * |
+ * [request] - The metadata request object. |
+ * |
+ * Request parameters: |
+ * |
+ * [variantId] - The ID of the variant to be updated. |
+ * |
+ * Completes with a [Variant]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<Variant> update(Variant request, core.String variantId) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (request != null) { |
+ _body = convert.JSON.encode((request).toJson()); |
+ } |
+ if (variantId == null) { |
+ throw new core.ArgumentError("Parameter variantId is required."); |
+ } |
+ |
+ |
+ _url = 'variants/' + common_internal.Escaper.ecapeVariable('$variantId'); |
+ |
+ var _response = _requester.request(_url, |
+ "PUT", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new Variant.fromJson(data)); |
+ } |
+ |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class VariantsetsResourceApi { |
+ final common_internal.ApiRequester _requester; |
+ |
+ VariantsetsResourceApi(common_internal.ApiRequester client) : |
+ _requester = client; |
+ |
+ /** |
+ * Deletes the contents of a variant set. The variant set object is not |
+ * deleted. |
+ * |
+ * Request parameters: |
+ * |
+ * [variantSetId] - The ID of the variant set to be deleted. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future delete(core.String variantSetId) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (variantSetId == null) { |
+ throw new core.ArgumentError("Parameter variantSetId is required."); |
+ } |
+ |
+ _downloadOptions = null; |
+ |
+ _url = 'variantsets/' + common_internal.Escaper.ecapeVariable('$variantSetId'); |
+ |
+ var _response = _requester.request(_url, |
+ "DELETE", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => null); |
+ } |
+ |
+ /** |
+ * Exports variant set data to an external destination. |
+ * |
+ * [request] - The metadata request object. |
+ * |
+ * Request parameters: |
+ * |
+ * [variantSetId] - Required. The ID of the variant set that contains variant |
+ * data which should be exported. The caller must have READ access to this |
+ * variant set. |
+ * |
+ * Completes with a [ExportVariantSetResponse]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<ExportVariantSetResponse> export(ExportVariantSetRequest request, core.String variantSetId) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (request != null) { |
+ _body = convert.JSON.encode((request).toJson()); |
+ } |
+ if (variantSetId == null) { |
+ throw new core.ArgumentError("Parameter variantSetId is required."); |
+ } |
+ |
+ |
+ _url = 'variantsets/' + common_internal.Escaper.ecapeVariable('$variantSetId') + '/export'; |
+ |
+ var _response = _requester.request(_url, |
+ "POST", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new ExportVariantSetResponse.fromJson(data)); |
+ } |
+ |
+ /** |
+ * Gets a variant set by ID. |
+ * |
+ * Request parameters: |
+ * |
+ * [variantSetId] - Required. The ID of the variant set. |
+ * |
+ * Completes with a [VariantSet]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<VariantSet> get(core.String variantSetId) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (variantSetId == null) { |
throw new core.ArgumentError("Parameter variantSetId is required."); |
} |
@@ -1734,6 +1998,60 @@ class VariantsetsResourceApi { |
} |
/** |
+ * Creates variant data by asynchronously importing the provided information. |
+ * |
+ * The variants for import will be merged with any existing data and each |
+ * other according to the behavior of mergeVariants. In particular, this means |
+ * for merged VCF variants that have conflicting INFO fields, some data will |
+ * be arbitrarily discarded. As a special case, for single-sample VCF files, |
+ * QUAL and FILTER fields will be moved to the call level; these are sometimes |
+ * interpreted in a call-specific context. Imported VCF headers are appended |
+ * to the metadata already in a variant set. |
+ * |
+ * [request] - The metadata request object. |
+ * |
+ * Request parameters: |
+ * |
+ * [variantSetId] - Required. The variant set to which variant data should be |
+ * imported. |
+ * |
+ * Completes with a [ImportVariantsResponse]. |
+ * |
+ * Completes with a [common.ApiRequestError] if the API endpoint returned an |
+ * error. |
+ * |
+ * If the used [http.Client] completes with an error when making a REST call, |
+ * this method will complete with the same error. |
+ */ |
+ async.Future<ImportVariantsResponse> importVariants(ImportVariantsRequest request, core.String variantSetId) { |
+ var _url = null; |
+ var _queryParams = new core.Map(); |
+ var _uploadMedia = null; |
+ var _uploadOptions = null; |
+ var _downloadOptions = common.DownloadOptions.Metadata; |
+ var _body = null; |
+ |
+ if (request != null) { |
+ _body = convert.JSON.encode((request).toJson()); |
+ } |
+ if (variantSetId == null) { |
+ throw new core.ArgumentError("Parameter variantSetId is required."); |
+ } |
+ |
+ |
+ _url = 'variantsets/' + common_internal.Escaper.ecapeVariable('$variantSetId') + '/importVariants'; |
+ |
+ var _response = _requester.request(_url, |
+ "POST", |
+ body: _body, |
+ queryParams: _queryParams, |
+ uploadOptions: _uploadOptions, |
+ uploadMedia: _uploadMedia, |
+ downloadOptions: _downloadOptions); |
+ return _response.then((data) => new ImportVariantsResponse.fromJson(data)); |
+ } |
+ |
+ /** |
* Merges the given variants with existing variants. Each variant will be |
* merged with an existing variant that matches its reference sequence, start, |
* end, reference bases, and alternative bases. If no such variant exists, a |
@@ -1786,8 +2104,7 @@ class VariantsetsResourceApi { |
/** |
* Updates a variant set's metadata. All other modifications are silently |
- * ignored. Returns the modified variant set. This method supports patch |
- * semantics. |
+ * ignored. This method supports patch semantics. |
* |
* [request] - The metadata request object. |
* |
@@ -1834,6 +2151,8 @@ class VariantsetsResourceApi { |
/** |
* Returns a list of all variant sets matching search criteria. |
* |
+ * Implements GlobalAllianceApi.searchVariantSets. |
+ * |
* [request] - The metadata request object. |
* |
* Request parameters: |
@@ -1873,7 +2192,7 @@ class VariantsetsResourceApi { |
/** |
* Updates a variant set's metadata. All other modifications are silently |
- * ignored. Returns the modified variant set. |
+ * ignored. |
* |
* [request] - The metadata request object. |
* |
@@ -1921,27 +2240,107 @@ class VariantsetsResourceApi { |
-/** |
- * A beacon represents whether any variant call in a variant set has a specific |
- * allele at a particular position. |
- */ |
-class Beacon { |
- /** True if the allele exists on any variant call, false otherwise. */ |
- core.bool exists; |
+/** The read group set align request. */ |
+class AlignReadGroupSetsRequest { |
+ /** |
+ * The BAM source files for alignment. Exactly one of readGroupSetIds, |
+ * bamSourceUris, interleavedFastqSource or pairedFastqSource must be |
+ * provided. The caller must have READ permissions for these files. |
+ */ |
+ core.List<core.String> bamSourceUris; |
+ /** |
+ * Required. The ID of the dataset the newly aligned read group sets will |
+ * belong to. The caller must have WRITE permissions to this dataset. |
+ */ |
+ core.String datasetId; |
- Beacon(); |
+ /** |
+ * The interleaved FASTQ source files for alignment, where both members of |
+ * each pair of reads are found on consecutive records within the same FASTQ |
+ * file. Exactly one of readGroupSetIds, bamSourceUris, interleavedFastqSource |
+ * or pairedFastqSource must be provided. |
+ */ |
+ InterleavedFastqSource interleavedFastqSource; |
+ |
+ /** |
+ * The paired end FASTQ source files for alignment, where each member of a |
+ * pair of reads are found in separate files. Exactly one of readGroupSetIds, |
+ * bamSourceUris, interleavedFastqSource or pairedFastqSource must be |
+ * provided. |
+ */ |
+ PairedFastqSource pairedFastqSource; |
+ |
+ /** |
+ * The IDs of the read group sets which will be aligned. New read group sets |
+ * will be generated to hold the aligned data, the originals will not be |
+ * modified. The caller must have READ permissions for these read group sets. |
+ * Exactly one of readGroupSetIds, bamSourceUris, interleavedFastqSource or |
+ * pairedFastqSource must be provided. |
+ */ |
+ core.List<core.String> readGroupSetIds; |
+ |
+ |
+ AlignReadGroupSetsRequest(); |
- Beacon.fromJson(core.Map _json) { |
- if (_json.containsKey("exists")) { |
- exists = _json["exists"]; |
+ AlignReadGroupSetsRequest.fromJson(core.Map _json) { |
+ if (_json.containsKey("bamSourceUris")) { |
+ bamSourceUris = _json["bamSourceUris"]; |
+ } |
+ if (_json.containsKey("datasetId")) { |
+ datasetId = _json["datasetId"]; |
+ } |
+ if (_json.containsKey("interleavedFastqSource")) { |
+ interleavedFastqSource = new InterleavedFastqSource.fromJson(_json["interleavedFastqSource"]); |
+ } |
+ if (_json.containsKey("pairedFastqSource")) { |
+ pairedFastqSource = new PairedFastqSource.fromJson(_json["pairedFastqSource"]); |
+ } |
+ if (_json.containsKey("readGroupSetIds")) { |
+ readGroupSetIds = _json["readGroupSetIds"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (exists != null) { |
- _json["exists"] = exists; |
+ if (bamSourceUris != null) { |
+ _json["bamSourceUris"] = bamSourceUris; |
+ } |
+ if (datasetId != null) { |
+ _json["datasetId"] = datasetId; |
+ } |
+ if (interleavedFastqSource != null) { |
+ _json["interleavedFastqSource"] = (interleavedFastqSource).toJson(); |
+ } |
+ if (pairedFastqSource != null) { |
+ _json["pairedFastqSource"] = (pairedFastqSource).toJson(); |
+ } |
+ if (readGroupSetIds != null) { |
+ _json["readGroupSetIds"] = readGroupSetIds; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** The read group set align response. */ |
+class AlignReadGroupSetsResponse { |
+ /** A job ID that can be used to get status information. */ |
+ core.String jobId; |
+ |
+ |
+ AlignReadGroupSetsResponse(); |
+ |
+ AlignReadGroupSetsResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("jobId")) { |
+ jobId = _json["jobId"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (jobId != null) { |
+ _json["jobId"] = jobId; |
} |
return _json; |
} |
@@ -1949,9 +2348,9 @@ class Beacon { |
/** |
- * A Call represents the determination of genotype with respect to a particular |
+ * A call represents the determination of genotype with respect to a particular |
* variant. It may include associated information such as quality and phasing. |
- * For example, a Call might assign a probability of 0.32 to the occurrence of a |
+ * For example, a call might assign a probability of 0.32 to the occurrence of a |
* SNP named rs1234 in a call set with the name NA12345. |
*/ |
class Call { |
@@ -2041,7 +2440,87 @@ class Call { |
} |
-/** A CallSet is a collection of variant calls. It belongs to a variant set. */ |
+/** The read group set call request. */ |
+class CallReadGroupSetsRequest { |
+ /** |
+ * Required. The ID of the dataset the called variants will belong to. The |
+ * caller must have WRITE permissions to this dataset. |
+ */ |
+ core.String datasetId; |
+ |
+ /** |
+ * The IDs of the read group sets which will be called. The caller must have |
+ * READ permissions for these read group sets. One of readGroupSetIds or |
+ * sourceUris must be provided. |
+ */ |
+ core.List<core.String> readGroupSetIds; |
+ |
+ /** |
+ * A list of URIs pointing at BAM files in Google Cloud Storage which will be |
+ * called. FASTQ files are not allowed. The caller must have READ permissions |
+ * for these files. One of readGroupSetIds or sourceUris must be provided. |
+ */ |
+ core.List<core.String> sourceUris; |
+ |
+ |
+ CallReadGroupSetsRequest(); |
+ |
+ CallReadGroupSetsRequest.fromJson(core.Map _json) { |
+ if (_json.containsKey("datasetId")) { |
+ datasetId = _json["datasetId"]; |
+ } |
+ if (_json.containsKey("readGroupSetIds")) { |
+ readGroupSetIds = _json["readGroupSetIds"]; |
+ } |
+ if (_json.containsKey("sourceUris")) { |
+ sourceUris = _json["sourceUris"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (datasetId != null) { |
+ _json["datasetId"] = datasetId; |
+ } |
+ if (readGroupSetIds != null) { |
+ _json["readGroupSetIds"] = readGroupSetIds; |
+ } |
+ if (sourceUris != null) { |
+ _json["sourceUris"] = sourceUris; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** The read group set call response. */ |
+class CallReadGroupSetsResponse { |
+ /** A job ID that can be used to get status information. */ |
+ core.String jobId; |
+ |
+ |
+ CallReadGroupSetsResponse(); |
+ |
+ CallReadGroupSetsResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("jobId")) { |
+ jobId = _json["jobId"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (jobId != null) { |
+ _json["jobId"] = jobId; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** |
+ * A call set is a collection of variant calls, typically for one sample. It |
+ * belongs to a variant set. |
+ */ |
class CallSet { |
/** The date this call set was created in milliseconds from the epoch. */ |
core.String created; |
@@ -2049,10 +2528,10 @@ class CallSet { |
/** The Google generated ID of the call set, immutable. */ |
core.String id; |
- /** A map of additional callset information. */ |
+ /** A map of additional call set information. */ |
core.Map<core.String, core.List<core.String>> info; |
- /** The callset name. */ |
+ /** The call set name. */ |
core.String name; |
/** The sample ID this call set corresponds to. */ |
@@ -2110,6 +2589,64 @@ class CallSet { |
} |
+/** A single CIGAR operation. */ |
+class CigarUnit { |
+ /** |
+ * Not documented yet. |
+ * Possible string values are: |
+ * - "ALIGNMENT_MATCH" |
+ * - "CLIP_HARD" |
+ * - "CLIP_SOFT" |
+ * - "DELETE" |
+ * - "INSERT" |
+ * - "PAD" |
+ * - "SEQUENCE_MATCH" |
+ * - "SEQUENCE_MISMATCH" |
+ * - "SKIP" |
+ */ |
+ core.String operation; |
+ |
+ /** The number of bases that the operation runs for. Required. */ |
+ core.String operationLength; |
+ |
+ /** |
+ * referenceSequence is only used at mismatches (SEQUENCE_MISMATCH) and |
+ * deletions (DELETE). Filling this field replaces SAM's MD tag. If the |
+ * relevant information is not available, this field is unset. |
+ */ |
+ core.String referenceSequence; |
+ |
+ |
+ CigarUnit(); |
+ |
+ CigarUnit.fromJson(core.Map _json) { |
+ if (_json.containsKey("operation")) { |
+ operation = _json["operation"]; |
+ } |
+ if (_json.containsKey("operationLength")) { |
+ operationLength = _json["operationLength"]; |
+ } |
+ if (_json.containsKey("referenceSequence")) { |
+ referenceSequence = _json["referenceSequence"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (operation != null) { |
+ _json["operation"] = operation; |
+ } |
+ if (operationLength != null) { |
+ _json["operationLength"] = operationLength; |
+ } |
+ if (referenceSequence != null) { |
+ _json["referenceSequence"] = referenceSequence; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
/** |
* A bucket over which read coverage has been precomputed. A bucket corresponds |
* to a specific range of the reference sequence. |
@@ -2122,7 +2659,7 @@ class CoverageBucket { |
core.double meanCoverage; |
/** The genomic coordinate range spanned by this bucket. */ |
- GenomicRange range; |
+ Range range; |
CoverageBucket(); |
@@ -2132,7 +2669,7 @@ class CoverageBucket { |
meanCoverage = _json["meanCoverage"]; |
} |
if (_json.containsKey("range")) { |
- range = new GenomicRange.fromJson(_json["range"]); |
+ range = new Range.fromJson(_json["range"]); |
} |
} |
@@ -2166,7 +2703,7 @@ class Dataset { |
/** |
* The Google Developers Console project number that this dataset belongs to. |
*/ |
- core.String projectId; |
+ core.String projectNumber; |
Dataset(); |
@@ -2181,8 +2718,8 @@ class Dataset { |
if (_json.containsKey("name")) { |
name = _json["name"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
} |
@@ -2197,8 +2734,8 @@ class Dataset { |
if (name != null) { |
_json["name"] = name; |
} |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
return _json; |
} |
@@ -2208,16 +2745,14 @@ class Dataset { |
/** The job creation request. */ |
class ExperimentalCreateJobRequest { |
/** |
- * Specifies whether or not to run the alignment pipeline. At least one of |
- * align or callVariants must be provided. |
+ * Specifies whether or not to run the alignment pipeline. Either align or |
+ * callVariants must be set. |
*/ |
core.bool align; |
/** |
- * Specifies whether or not to run the variant calling pipeline. If specified, |
- * alignment will be performed first and the aligned BAMs will passed as input |
- * to the variant caller. At least one of align or callVariants must be |
- * provided. |
+ * Specifies whether or not to run the variant calling pipeline. Either align |
+ * or callVariants must be set. |
*/ |
core.bool callVariants; |
@@ -2237,7 +2772,7 @@ class ExperimentalCreateJobRequest { |
/** |
* Required. The Google Cloud Project ID with which to associate the request. |
*/ |
- core.String projectId; |
+ core.String projectNumber; |
/** |
* A list of Google Cloud Storage URIs of data files to operate upon. These |
@@ -2263,8 +2798,8 @@ class ExperimentalCreateJobRequest { |
if (_json.containsKey("pairedSourceUris")) { |
pairedSourceUris = _json["pairedSourceUris"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
if (_json.containsKey("sourceUris")) { |
sourceUris = _json["sourceUris"]; |
@@ -2285,8 +2820,8 @@ class ExperimentalCreateJobRequest { |
if (pairedSourceUris != null) { |
_json["pairedSourceUris"] = pairedSourceUris; |
} |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
if (sourceUris != null) { |
_json["sourceUris"] = sourceUris; |
@@ -2320,8 +2855,8 @@ class ExperimentalCreateJobResponse { |
} |
-/** The readset export request. */ |
-class ExportReadsetsRequest { |
+/** The read group set export request. */ |
+class ExportReadGroupSetsRequest { |
/** |
* A Google Cloud Storage URI where the exported BAM file will be created. The |
* currently authenticated user must have write access to the new file |
@@ -2330,10 +2865,10 @@ class ExportReadsetsRequest { |
core.String exportUri; |
/** The Google Developers Console project number that owns this export. */ |
- core.String projectId; |
+ core.String projectNumber; |
- /** The IDs of the readsets to export. */ |
- core.List<core.String> readsetIds; |
+ /** The IDs of the read group sets to export. */ |
+ core.List<core.String> readGroupSetIds; |
/** |
* The reference names to export. If this is not specified, all reference |
@@ -2343,17 +2878,17 @@ class ExportReadsetsRequest { |
core.List<core.String> referenceNames; |
- ExportReadsetsRequest(); |
+ ExportReadGroupSetsRequest(); |
- ExportReadsetsRequest.fromJson(core.Map _json) { |
+ ExportReadGroupSetsRequest.fromJson(core.Map _json) { |
if (_json.containsKey("exportUri")) { |
exportUri = _json["exportUri"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
- if (_json.containsKey("readsetIds")) { |
- readsetIds = _json["readsetIds"]; |
+ if (_json.containsKey("readGroupSetIds")) { |
+ readGroupSetIds = _json["readGroupSetIds"]; |
} |
if (_json.containsKey("referenceNames")) { |
referenceNames = _json["referenceNames"]; |
@@ -2365,11 +2900,11 @@ class ExportReadsetsRequest { |
if (exportUri != null) { |
_json["exportUri"] = exportUri; |
} |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
- if (readsetIds != null) { |
- _json["readsetIds"] = readsetIds; |
+ if (readGroupSetIds != null) { |
+ _json["readGroupSetIds"] = readGroupSetIds; |
} |
if (referenceNames != null) { |
_json["referenceNames"] = referenceNames; |
@@ -2379,15 +2914,15 @@ class ExportReadsetsRequest { |
} |
-/** The readset export response. */ |
-class ExportReadsetsResponse { |
+/** The read group set export response. */ |
+class ExportReadGroupSetsResponse { |
/** A job ID that can be used to get status information. */ |
core.String jobId; |
- ExportReadsetsResponse(); |
+ ExportReadGroupSetsResponse(); |
- ExportReadsetsResponse.fromJson(core.Map _json) { |
+ ExportReadGroupSetsResponse.fromJson(core.Map _json) { |
if (_json.containsKey("jobId")) { |
jobId = _json["jobId"]; |
} |
@@ -2404,7 +2939,7 @@ class ExportReadsetsResponse { |
/** The variant data export request. */ |
-class ExportVariantsRequest { |
+class ExportVariantSetRequest { |
/** |
* The BigQuery dataset to export data to. Note that this is distinct from the |
* Genomics concept of "dataset". |
@@ -2426,7 +2961,7 @@ class ExportVariantsRequest { |
/** |
* The format for the exported data. |
* Possible string values are: |
- * - "bigquery" |
+ * - "BIGQUERY" |
*/ |
core.String format; |
@@ -2435,18 +2970,12 @@ class ExportVariantsRequest { |
* The caller must have WRITE access to this project. This project will also |
* own the resulting export job. |
*/ |
- core.String projectId; |
- |
- /** |
- * Required. The ID of the variant set that contains variant data which should |
- * be exported. The caller must have READ access to this variant set. |
- */ |
- core.String variantSetId; |
+ core.String projectNumber; |
- ExportVariantsRequest(); |
+ ExportVariantSetRequest(); |
- ExportVariantsRequest.fromJson(core.Map _json) { |
+ ExportVariantSetRequest.fromJson(core.Map _json) { |
if (_json.containsKey("bigqueryDataset")) { |
bigqueryDataset = _json["bigqueryDataset"]; |
} |
@@ -2459,11 +2988,8 @@ class ExportVariantsRequest { |
if (_json.containsKey("format")) { |
format = _json["format"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
- } |
- if (_json.containsKey("variantSetId")) { |
- variantSetId = _json["variantSetId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
} |
@@ -2475,229 +3001,139 @@ class ExportVariantsRequest { |
if (bigqueryTable != null) { |
_json["bigqueryTable"] = bigqueryTable; |
} |
- if (callSetIds != null) { |
- _json["callSetIds"] = callSetIds; |
- } |
- if (format != null) { |
- _json["format"] = format; |
- } |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
- } |
- if (variantSetId != null) { |
- _json["variantSetId"] = variantSetId; |
- } |
- return _json; |
- } |
-} |
- |
- |
-/** The variant data export response. */ |
-class ExportVariantsResponse { |
- /** A job ID that can be used to get status information. */ |
- core.String jobId; |
- |
- |
- ExportVariantsResponse(); |
- |
- ExportVariantsResponse.fromJson(core.Map _json) { |
- if (_json.containsKey("jobId")) { |
- jobId = _json["jobId"]; |
- } |
- } |
- |
- core.Map toJson() { |
- var _json = new core.Map(); |
- if (jobId != null) { |
- _json["jobId"] = jobId; |
- } |
- return _json; |
- } |
-} |
- |
- |
-/** |
- * An inclusive, exclusive genomic coordinate range over a reference sequence. |
- */ |
-class GenomicRange { |
- /** |
- * The end position of the range on the reference, 1-based exclusive. If |
- * specified, sequenceName must also be specified. |
- */ |
- core.String sequenceEnd; |
- |
- /** The reference sequence name, for example chr1, 1, or chrX. */ |
- core.String sequenceName; |
- |
- /** |
- * The start position of the range on the reference, 1-based inclusive. If |
- * specified, sequenceName must also be specified. |
- */ |
- core.String sequenceStart; |
- |
- |
- GenomicRange(); |
- |
- GenomicRange.fromJson(core.Map _json) { |
- if (_json.containsKey("sequenceEnd")) { |
- sequenceEnd = _json["sequenceEnd"]; |
- } |
- if (_json.containsKey("sequenceName")) { |
- sequenceName = _json["sequenceName"]; |
- } |
- if (_json.containsKey("sequenceStart")) { |
- sequenceStart = _json["sequenceStart"]; |
- } |
- } |
- |
- core.Map toJson() { |
- var _json = new core.Map(); |
- if (sequenceEnd != null) { |
- _json["sequenceEnd"] = sequenceEnd; |
+ if (callSetIds != null) { |
+ _json["callSetIds"] = callSetIds; |
} |
- if (sequenceName != null) { |
- _json["sequenceName"] = sequenceName; |
+ if (format != null) { |
+ _json["format"] = format; |
} |
- if (sequenceStart != null) { |
- _json["sequenceStart"] = sequenceStart; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
return _json; |
} |
} |
-/** Not documented yet. */ |
-class Header { |
- /** (SO) Sorting order of alignments. */ |
- core.String sortingOrder; |
- |
- /** (VN) BAM format version. */ |
- core.String version; |
+/** The variant data export response. */ |
+class ExportVariantSetResponse { |
+ /** A job ID that can be used to get status information. */ |
+ core.String jobId; |
- Header(); |
+ ExportVariantSetResponse(); |
- Header.fromJson(core.Map _json) { |
- if (_json.containsKey("sortingOrder")) { |
- sortingOrder = _json["sortingOrder"]; |
- } |
- if (_json.containsKey("version")) { |
- version = _json["version"]; |
+ ExportVariantSetResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("jobId")) { |
+ jobId = _json["jobId"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (sortingOrder != null) { |
- _json["sortingOrder"] = sortingOrder; |
- } |
- if (version != null) { |
- _json["version"] = version; |
+ if (jobId != null) { |
+ _json["jobId"] = jobId; |
} |
return _json; |
} |
} |
-/** The header section of the BAM/SAM file. */ |
-class HeaderSection { |
- /** (@CO) One-line text comments. */ |
- core.List<core.String> comments; |
+/** Not documented yet. */ |
+class FastqMetadata { |
+ /** Optionally specifies the library name for alignment from FASTQ. */ |
+ core.String libraryName; |
/** |
- * [Deprecated] This field is deprecated and will no longer be populated. |
- * Please use filename instead. |
+ * Optionally specifies the platform name for alignment from FASTQ. For |
+ * example: CAPILLARY, LS454, ILLUMINA, SOLID, HELICOS, IONTORRENT, PACBIO. |
*/ |
- core.String fileUri; |
- |
- /** The name of the file from which this data was imported. */ |
- core.String filename; |
- |
- /** (@HD) The header line. */ |
- core.List<Header> headers; |
+ core.String platformName; |
- /** (@PG) Programs. */ |
- core.List<Program> programs; |
+ /** |
+ * Optionally specifies the platform unit for alignment from FASTQ. For |
+ * example: flowcell-barcode.lane for Illumina or slide for SOLID. |
+ */ |
+ core.String platformUnit; |
- /** (@RG) Read group. */ |
- core.List<ReadGroup> readGroups; |
+ /** Optionally specifies the read group name for alignment from FASTQ. */ |
+ core.String readGroupName; |
- /** (@SQ) Reference sequence dictionary. */ |
- core.List<ReferenceSequence> refSequences; |
+ /** Optionally specifies the sample name for alignment from FASTQ. */ |
+ core.String sampleName; |
- HeaderSection(); |
+ FastqMetadata(); |
- HeaderSection.fromJson(core.Map _json) { |
- if (_json.containsKey("comments")) { |
- comments = _json["comments"]; |
+ FastqMetadata.fromJson(core.Map _json) { |
+ if (_json.containsKey("libraryName")) { |
+ libraryName = _json["libraryName"]; |
} |
- if (_json.containsKey("fileUri")) { |
- fileUri = _json["fileUri"]; |
+ if (_json.containsKey("platformName")) { |
+ platformName = _json["platformName"]; |
} |
- if (_json.containsKey("filename")) { |
- filename = _json["filename"]; |
- } |
- if (_json.containsKey("headers")) { |
- headers = _json["headers"].map((value) => new Header.fromJson(value)).toList(); |
- } |
- if (_json.containsKey("programs")) { |
- programs = _json["programs"].map((value) => new Program.fromJson(value)).toList(); |
+ if (_json.containsKey("platformUnit")) { |
+ platformUnit = _json["platformUnit"]; |
} |
- if (_json.containsKey("readGroups")) { |
- readGroups = _json["readGroups"].map((value) => new ReadGroup.fromJson(value)).toList(); |
+ if (_json.containsKey("readGroupName")) { |
+ readGroupName = _json["readGroupName"]; |
} |
- if (_json.containsKey("refSequences")) { |
- refSequences = _json["refSequences"].map((value) => new ReferenceSequence.fromJson(value)).toList(); |
+ if (_json.containsKey("sampleName")) { |
+ sampleName = _json["sampleName"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (comments != null) { |
- _json["comments"] = comments; |
- } |
- if (fileUri != null) { |
- _json["fileUri"] = fileUri; |
- } |
- if (filename != null) { |
- _json["filename"] = filename; |
+ if (libraryName != null) { |
+ _json["libraryName"] = libraryName; |
} |
- if (headers != null) { |
- _json["headers"] = headers.map((value) => (value).toJson()).toList(); |
+ if (platformName != null) { |
+ _json["platformName"] = platformName; |
} |
- if (programs != null) { |
- _json["programs"] = programs.map((value) => (value).toJson()).toList(); |
+ if (platformUnit != null) { |
+ _json["platformUnit"] = platformUnit; |
} |
- if (readGroups != null) { |
- _json["readGroups"] = readGroups.map((value) => (value).toJson()).toList(); |
+ if (readGroupName != null) { |
+ _json["readGroupName"] = readGroupName; |
} |
- if (refSequences != null) { |
- _json["refSequences"] = refSequences.map((value) => (value).toJson()).toList(); |
+ if (sampleName != null) { |
+ _json["sampleName"] = sampleName; |
} |
return _json; |
} |
} |
-/** The readset import request. */ |
-class ImportReadsetsRequest { |
+/** The read group set import request. */ |
+class ImportReadGroupSetsRequest { |
/** |
- * Required. The ID of the dataset these readsets will belong to. The caller |
- * must have WRITE permissions to this dataset. |
+ * Required. The ID of the dataset these read group sets will belong to. The |
+ * caller must have WRITE permissions to this dataset. |
*/ |
core.String datasetId; |
+ /** |
+ * The reference set to which the imported read group sets are aligned to, if |
+ * any. The reference names of this reference set must be a superset of those |
+ * found in the imported file headers. If no reference set id is provided, a |
+ * best effort is made to associate with a matching reference set. |
+ */ |
+ core.String referenceSetId; |
+ |
/** A list of URIs pointing at BAM files in Google Cloud Storage. */ |
core.List<core.String> sourceUris; |
- ImportReadsetsRequest(); |
+ ImportReadGroupSetsRequest(); |
- ImportReadsetsRequest.fromJson(core.Map _json) { |
+ ImportReadGroupSetsRequest.fromJson(core.Map _json) { |
if (_json.containsKey("datasetId")) { |
datasetId = _json["datasetId"]; |
} |
+ if (_json.containsKey("referenceSetId")) { |
+ referenceSetId = _json["referenceSetId"]; |
+ } |
if (_json.containsKey("sourceUris")) { |
sourceUris = _json["sourceUris"]; |
} |
@@ -2708,6 +3144,9 @@ class ImportReadsetsRequest { |
if (datasetId != null) { |
_json["datasetId"] = datasetId; |
} |
+ if (referenceSetId != null) { |
+ _json["referenceSetId"] = referenceSetId; |
+ } |
if (sourceUris != null) { |
_json["sourceUris"] = sourceUris; |
} |
@@ -2716,15 +3155,15 @@ class ImportReadsetsRequest { |
} |
-/** The readset import response. */ |
-class ImportReadsetsResponse { |
+/** The read group set import response. */ |
+class ImportReadGroupSetsResponse { |
/** A job ID that can be used to get status information. */ |
core.String jobId; |
- ImportReadsetsResponse(); |
+ ImportReadGroupSetsResponse(); |
- ImportReadsetsResponse.fromJson(core.Map _json) { |
+ ImportReadGroupSetsResponse.fromJson(core.Map _json) { |
if (_json.containsKey("jobId")) { |
jobId = _json["jobId"]; |
} |
@@ -2745,8 +3184,8 @@ class ImportVariantsRequest { |
/** |
* The format of the variant data being imported. |
* Possible string values are: |
- * - "completeGenomics" |
- * - "vcf" |
+ * - "COMPLETE_GENOMICS" |
+ * - "VCF" |
*/ |
core.String format; |
@@ -2756,9 +3195,6 @@ class ImportVariantsRequest { |
*/ |
core.List<core.String> sourceUris; |
- /** Required. The variant set to which variant data should be imported. */ |
- core.String variantSetId; |
- |
ImportVariantsRequest(); |
@@ -2769,9 +3205,6 @@ class ImportVariantsRequest { |
if (_json.containsKey("sourceUris")) { |
sourceUris = _json["sourceUris"]; |
} |
- if (_json.containsKey("variantSetId")) { |
- variantSetId = _json["variantSetId"]; |
- } |
} |
core.Map toJson() { |
@@ -2782,9 +3215,6 @@ class ImportVariantsRequest { |
if (sourceUris != null) { |
_json["sourceUris"] = sourceUris; |
} |
- if (variantSetId != null) { |
- _json["variantSetId"] = variantSetId; |
- } |
return _json; |
} |
} |
@@ -2814,6 +3244,46 @@ class ImportVariantsResponse { |
} |
+/** Describes an interleaved FASTQ file source for alignment. */ |
+class InterleavedFastqSource { |
+ /** |
+ * Optionally specifies the metadata to be associated with the final aligned |
+ * read group set. |
+ */ |
+ FastqMetadata metadata; |
+ |
+ /** |
+ * A list of URIs pointing at interleaved FASTQ files in Google Cloud Storage |
+ * which will be aligned. The caller must have READ permissions for these |
+ * files. |
+ */ |
+ core.List<core.String> sourceUris; |
+ |
+ |
+ InterleavedFastqSource(); |
+ |
+ InterleavedFastqSource.fromJson(core.Map _json) { |
+ if (_json.containsKey("metadata")) { |
+ metadata = new FastqMetadata.fromJson(_json["metadata"]); |
+ } |
+ if (_json.containsKey("sourceUris")) { |
+ sourceUris = _json["sourceUris"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (metadata != null) { |
+ _json["metadata"] = (metadata).toJson(); |
+ } |
+ if (sourceUris != null) { |
+ _json["sourceUris"] = sourceUris; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
/** |
* A Job represents an ongoing process that can be monitored for status |
* information. |
@@ -2823,7 +3293,7 @@ class Job { |
core.String created; |
/** A more detailed description of this job's current status. */ |
- core.String description; |
+ core.String detailedStatus; |
/** Any errors that occurred during processing. */ |
core.List<core.String> errors; |
@@ -2840,7 +3310,7 @@ class Job { |
/** |
* The Google Developers Console project number to which this job belongs. |
*/ |
- core.String projectId; |
+ core.String projectNumber; |
/** A summarized representation of the original service request. */ |
JobRequest request; |
@@ -2848,13 +3318,13 @@ class Job { |
/** |
* The status of this job. |
* Possible string values are: |
- * - "canceled" |
- * - "failure" |
- * - "new" |
- * - "pending" |
- * - "running" |
- * - "success" |
- * - "unknownStatus" |
+ * - "CANCELED" |
+ * - "FAILURE" |
+ * - "NEW" |
+ * - "PENDING" |
+ * - "RUNNING" |
+ * - "SUCCESS" |
+ * - "UNKNOWN_STATUS" |
*/ |
core.String status; |
@@ -2868,8 +3338,8 @@ class Job { |
if (_json.containsKey("created")) { |
created = _json["created"]; |
} |
- if (_json.containsKey("description")) { |
- description = _json["description"]; |
+ if (_json.containsKey("detailedStatus")) { |
+ detailedStatus = _json["detailedStatus"]; |
} |
if (_json.containsKey("errors")) { |
errors = _json["errors"]; |
@@ -2880,8 +3350,8 @@ class Job { |
if (_json.containsKey("importedIds")) { |
importedIds = _json["importedIds"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
if (_json.containsKey("request")) { |
request = new JobRequest.fromJson(_json["request"]); |
@@ -2899,8 +3369,8 @@ class Job { |
if (created != null) { |
_json["created"] = created; |
} |
- if (description != null) { |
- _json["description"] = description; |
+ if (detailedStatus != null) { |
+ _json["detailedStatus"] = detailedStatus; |
} |
if (errors != null) { |
_json["errors"] = errors; |
@@ -2911,8 +3381,8 @@ class Job { |
if (importedIds != null) { |
_json["importedIds"] = importedIds; |
} |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
if (request != null) { |
_json["request"] = (request).toJson(); |
@@ -2945,12 +3415,14 @@ class JobRequest { |
/** |
* The original request type. |
* Possible string values are: |
- * - "experimentalCreateJob" |
- * - "exportReadsets" |
- * - "exportVariants" |
- * - "importReadsets" |
- * - "importVariants" |
- * - "unknownType" |
+ * - "ALIGN_READSETS" |
+ * - "CALL_READSETS" |
+ * - "EXPERIMENTAL_CREATE_JOB" |
+ * - "EXPORT_READSETS" |
+ * - "EXPORT_VARIANTS" |
+ * - "IMPORT_READSETS" |
+ * - "IMPORT_VARIANTS" |
+ * - "UNKNOWN_TYPE" |
*/ |
core.String type; |
@@ -2985,6 +3457,106 @@ class JobRequest { |
} |
+/** |
+ * A linear alignment can be represented by one CIGAR string. Describes the |
+ * mapped position and local alignment of the read to the reference. |
+ */ |
+class LinearAlignment { |
+ /** |
+ * Represents the local alignment of this sequence (alignment matches, indels, |
+ * etc) against the reference. |
+ */ |
+ core.List<CigarUnit> cigar; |
+ |
+ /** |
+ * The mapping quality of this alignment. Represents how likely the read maps |
+ * to this position as opposed to other locations. |
+ */ |
+ core.int mappingQuality; |
+ |
+ /** The position of this alignment. */ |
+ Position position; |
+ |
+ |
+ LinearAlignment(); |
+ |
+ LinearAlignment.fromJson(core.Map _json) { |
+ if (_json.containsKey("cigar")) { |
+ cigar = _json["cigar"].map((value) => new CigarUnit.fromJson(value)).toList(); |
+ } |
+ if (_json.containsKey("mappingQuality")) { |
+ mappingQuality = _json["mappingQuality"]; |
+ } |
+ if (_json.containsKey("position")) { |
+ position = new Position.fromJson(_json["position"]); |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (cigar != null) { |
+ _json["cigar"] = cigar.map((value) => (value).toJson()).toList(); |
+ } |
+ if (mappingQuality != null) { |
+ _json["mappingQuality"] = mappingQuality; |
+ } |
+ if (position != null) { |
+ _json["position"] = (position).toJson(); |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ListBasesResponse { |
+ /** |
+ * The continuation token, which is used to page through large result sets. |
+ * Provide this value in a subsequent request to return the next page of |
+ * results. This field will be empty if there aren't any additional results. |
+ */ |
+ core.String nextPageToken; |
+ |
+ /** |
+ * The offset position (0-based) of the given sequence from the start of this |
+ * Reference. This value will differ for each page in a paginated request. |
+ */ |
+ core.String offset; |
+ |
+ /** A substring of the bases that make up this reference. */ |
+ core.String sequence; |
+ |
+ |
+ ListBasesResponse(); |
+ |
+ ListBasesResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("nextPageToken")) { |
+ nextPageToken = _json["nextPageToken"]; |
+ } |
+ if (_json.containsKey("offset")) { |
+ offset = _json["offset"]; |
+ } |
+ if (_json.containsKey("sequence")) { |
+ sequence = _json["sequence"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (nextPageToken != null) { |
+ _json["nextPageToken"] = nextPageToken; |
+ } |
+ if (offset != null) { |
+ _json["offset"] = offset; |
+ } |
+ if (sequence != null) { |
+ _json["sequence"] = sequence; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
/** Not documented yet. */ |
class ListCoverageBucketsResponse { |
/** |
@@ -3134,12 +3706,12 @@ class Metadata { |
* The type of data. Possible types include: Integer, Float, Flag, Character, |
* and String. |
* Possible string values are: |
- * - "character" |
- * - "flag" |
- * - "float" |
- * - "integer" |
- * - "string" |
- * - "unknownType" |
+ * - "CHARACTER" |
+ * - "FLAG" |
+ * - "FLOAT" |
+ * - "INTEGER" |
+ * - "STRING" |
+ * - "UNKNOWN_TYPE" |
*/ |
core.String type; |
@@ -3201,60 +3773,159 @@ class Metadata { |
} |
-/** Not documented yet. */ |
-class Program { |
- /** (CL) Command line. */ |
- core.String commandLine; |
- |
- /** (ID) Program record identifier. */ |
- core.String id; |
+/** Describes a paired-end FASTQ file source for alignment. */ |
+class PairedFastqSource { |
+ /** |
+ * A list of URIs pointing at paired end FASTQ files in Google Cloud Storage |
+ * which will be aligned. The first of each paired file should be specified |
+ * here, in an order that matches the second of each paired file specified in |
+ * secondSourceUris. For example: firstSourceUris: [file1_1.fq, file2_1.fq], |
+ * secondSourceUris: [file1_2.fq, file2_2.fq]. The caller must have READ |
+ * permissions for these files. |
+ */ |
+ core.List<core.String> firstSourceUris; |
- /** (PN) Program name. */ |
- core.String name; |
+ /** |
+ * Optionally specifies the metadata to be associated with the final aligned |
+ * read group set. |
+ */ |
+ FastqMetadata metadata; |
- /** (PP) Previous program ID. */ |
- core.String prevProgramId; |
+ /** |
+ * A list of URIs pointing at paired end FASTQ files in Google Cloud Storage |
+ * which will be aligned. The second of each paired file should be specified |
+ * here, in an order that matches the first of each paired file specified in |
+ * firstSourceUris. For example: firstSourceUris: [file1_1.fq, file2_1.fq], |
+ * secondSourceUris: [file1_2.fq, file2_2.fq]. The caller must have READ |
+ * permissions for these files. |
+ */ |
+ core.List<core.String> secondSourceUris; |
- /** (VN) Program version. */ |
- core.String version; |
+ PairedFastqSource(); |
- Program(); |
+ PairedFastqSource.fromJson(core.Map _json) { |
+ if (_json.containsKey("firstSourceUris")) { |
+ firstSourceUris = _json["firstSourceUris"]; |
+ } |
+ if (_json.containsKey("metadata")) { |
+ metadata = new FastqMetadata.fromJson(_json["metadata"]); |
+ } |
+ if (_json.containsKey("secondSourceUris")) { |
+ secondSourceUris = _json["secondSourceUris"]; |
+ } |
+ } |
- Program.fromJson(core.Map _json) { |
- if (_json.containsKey("commandLine")) { |
- commandLine = _json["commandLine"]; |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (firstSourceUris != null) { |
+ _json["firstSourceUris"] = firstSourceUris; |
} |
- if (_json.containsKey("id")) { |
- id = _json["id"]; |
+ if (metadata != null) { |
+ _json["metadata"] = (metadata).toJson(); |
} |
- if (_json.containsKey("name")) { |
- name = _json["name"]; |
+ if (secondSourceUris != null) { |
+ _json["secondSourceUris"] = secondSourceUris; |
} |
- if (_json.containsKey("prevProgramId")) { |
- prevProgramId = _json["prevProgramId"]; |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** |
+ * An abstraction for referring to a genomic position, in relation to some |
+ * already known reference. For now, represents a genomic position as a |
+ * reference name, a base number on that reference (0-based), and a |
+ * determination of forward or reverse strand. |
+ */ |
+class Position { |
+ /** |
+ * The 0-based offset from the start of the forward strand for that reference. |
+ */ |
+ core.String position; |
+ |
+ /** The name of the reference in whatever reference set is being used. */ |
+ core.String referenceName; |
+ |
+ /** |
+ * Whether this position is on the reverse strand, as opposed to the forward |
+ * strand. |
+ */ |
+ core.bool reverseStrand; |
+ |
+ |
+ Position(); |
+ |
+ Position.fromJson(core.Map _json) { |
+ if (_json.containsKey("position")) { |
+ position = _json["position"]; |
} |
- if (_json.containsKey("version")) { |
- version = _json["version"]; |
+ if (_json.containsKey("referenceName")) { |
+ referenceName = _json["referenceName"]; |
+ } |
+ if (_json.containsKey("reverseStrand")) { |
+ reverseStrand = _json["reverseStrand"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (commandLine != null) { |
- _json["commandLine"] = commandLine; |
+ if (position != null) { |
+ _json["position"] = position; |
} |
- if (id != null) { |
- _json["id"] = id; |
+ if (referenceName != null) { |
+ _json["referenceName"] = referenceName; |
} |
- if (name != null) { |
- _json["name"] = name; |
+ if (reverseStrand != null) { |
+ _json["reverseStrand"] = reverseStrand; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** A 0-based half-open genomic coordinate range over a reference sequence. */ |
+class Range { |
+ /** |
+ * The end position of the range on the reference, 0-based exclusive. If |
+ * specified, referenceName must also be specified. |
+ */ |
+ core.String end; |
+ |
+ /** The reference sequence name, for example chr1, 1, or chrX. */ |
+ core.String referenceName; |
+ |
+ /** |
+ * The start position of the range on the reference, 0-based inclusive. If |
+ * specified, referenceName must also be specified. |
+ */ |
+ core.String start; |
+ |
+ |
+ Range(); |
+ |
+ Range.fromJson(core.Map _json) { |
+ if (_json.containsKey("end")) { |
+ end = _json["end"]; |
+ } |
+ if (_json.containsKey("referenceName")) { |
+ referenceName = _json["referenceName"]; |
+ } |
+ if (_json.containsKey("start")) { |
+ start = _json["start"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (end != null) { |
+ _json["end"] = end; |
} |
- if (prevProgramId != null) { |
- _json["prevProgramId"] = prevProgramId; |
+ if (referenceName != null) { |
+ _json["referenceName"] = referenceName; |
} |
- if (version != null) { |
- _json["version"] = version; |
+ if (start != null) { |
+ _json["start"] = start; |
} |
return _json; |
} |
@@ -3262,345 +3933,527 @@ class Program { |
/** |
- * A Read is a group of bases that may or may not have been aligned to a |
- * reference. It contains quality information and other metadata. |
+ * A read alignment describes a linear alignment of a string of DNA to a |
+ * reference sequence, in addition to metadata about the fragment (the molecule |
+ * of DNA sequenced) and the read (the bases which were read by the sequencer). |
+ * A read is equivalent to a line in a SAM file. A read belongs to exactly one |
+ * read group and exactly one read group set. |
*/ |
class Read { |
/** |
- * The originalBases after the cigar field has been applied. Deletions are |
- * represented with '-' and insertions are omitted. |
+ * The quality of the read sequence contained in this alignment record. |
+ * alignedSequence and alignedQuality may be shorter than the full read |
+ * sequence and quality. This will occur if the alignment is part of a |
+ * chimeric alignment, or if the read was trimmed. When this occurs, the CIGAR |
+ * for this read will begin/end with a hard clip operator that will indicate |
+ * the length of the excised sequence. |
*/ |
- core.String alignedBases; |
+ core.List<core.int> alignedQuality; |
/** |
- * Represents the quality of each base in this read. Each character represents |
- * one base. To get the quality, take the ASCII value of the character and |
- * subtract 33. (QUAL) |
+ * The bases of the read sequence contained in this alignment record. |
+ * alignedSequence and alignedQuality may be shorter than the full read |
+ * sequence and quality. This will occur if the alignment is part of a |
+ * chimeric alignment, or if the read was trimmed. When this occurs, the CIGAR |
+ * for this read will begin/end with a hard clip operator that will indicate |
+ * the length of the excised sequence. |
*/ |
- core.String baseQuality; |
+ core.String alignedSequence; |
/** |
- * A condensed representation of how this read matches up to the reference. |
- * (CIGAR) |
+ * The linear alignment for this alignment record. This field will be null if |
+ * the read is unmapped. |
*/ |
- core.String cigar; |
+ LinearAlignment alignment; |
+ |
+ /** The fragment is a PCR or optical duplicate (SAM flag 0x400) */ |
+ core.bool duplicateFragment; |
+ |
+ /** SAM flag 0x200 */ |
+ core.bool failedVendorQualityChecks; |
+ |
+ /** The observed length of the fragment, equivalent to TLEN in SAM. */ |
+ core.int fragmentLength; |
+ |
+ /** The fragment name. Equivalent to QNAME (query template name) in SAM. */ |
+ core.String fragmentName; |
/** |
- * Each bit of this number has a different meaning if enabled. See the full |
- * BAM spec for more details. (FLAG) |
+ * The unique ID for this read. This is a generated unique ID, not to be |
+ * confused with fragmentName. |
*/ |
- core.int flags; |
- |
- /** The Google generated ID of the read, immutable. */ |
core.String id; |
+ /** A map of additional read alignment information. */ |
+ core.Map<core.String, core.List<core.String>> info; |
+ |
/** |
- * A score up to 255 that represents how likely this read's aligned position |
- * is to be correct. A higher value is better. (MAPQ) |
+ * The mapping of the primary alignment of the (readNumber+1)%numberReads read |
+ * in the fragment. It replaces mate position and mate strand in SAM. |
*/ |
- core.int mappingQuality; |
+ Position nextMatePosition; |
- /** The 1-based start position of the paired read. (PNEXT) */ |
- core.int matePosition; |
+ /** The number of reads in the fragment (extension to SAM flag 0x1). */ |
+ core.int numberReads; |
/** |
- * The name of the sequence that the paired read is aligned to. This is |
- * usually the same as referenceSequenceName. (RNEXT) |
+ * The orientation and the distance between reads from the fragment are |
+ * consistent with the sequencing protocol (extension to SAM flag 0x2) |
*/ |
- core.String mateReferenceSequenceName; |
+ core.bool properPlacement; |
/** |
- * The name of the read. When imported from a BAM file, this is the query |
- * template name. (QNAME) |
+ * The ID of the read group this read belongs to. (Every read must belong to |
+ * exactly one read group.) |
*/ |
- core.String name; |
- |
- /** The list of bases that this read represents (such as "CATCGA"). (SEQ) */ |
- core.String originalBases; |
+ core.String readGroupId; |
/** |
- * The 1-based start position of the aligned read. If the first base starts at |
- * the very beginning of the reference sequence, then the position would be |
- * '1'. (POS) |
+ * The ID of the read group set this read belongs to. (Every read must belong |
+ * to exactly one read group set.) |
*/ |
- core.int position; |
- |
- /** The ID of the readset this read belongs to. */ |
- core.String readsetId; |
+ core.String readGroupSetId; |
/** |
- * The name of the sequence that this read is aligned to. This would be, for |
- * example, 'X' for the X Chromosome or '20' for Chromosome 20. (RNAME) |
+ * The read number in sequencing. 0-based and less than numberReads. This |
+ * field replaces SAM flag 0x40 and 0x80. |
*/ |
- core.String referenceSequenceName; |
+ core.int readNumber; |
- /** A map of additional read information. (TAG) */ |
- core.Map<core.String, core.List<core.String>> tags; |
+ /** |
+ * Whether this alignment is secondary. Equivalent to SAM flag 0x100. A |
+ * secondary alignment represents an alternative to the primary alignment for |
+ * this read. Aligners may return secondary alignments if a read can map |
+ * ambiguously to multiple coordinates in the genome. By convention, each read |
+ * has one and only one alignment where both secondaryAlignment and |
+ * supplementaryAlignment are false. |
+ */ |
+ core.bool secondaryAlignment; |
/** |
- * Length of the original piece of DNA that produced both this read and the |
- * paired read. (TLEN) |
+ * Whether this alignment is supplementary. Equivalent to SAM flag 0x800. |
+ * Supplementary alignments are used in the representation of a chimeric |
+ * alignment. In a chimeric alignment, a read is split into multiple linear |
+ * alignments that map to different reference contigs. The first linear |
+ * alignment in the read will be designated as the representative alignment; |
+ * the remaining linear alignments will be designated as supplementary |
+ * alignments. These alignments may have different mapping quality scores. In |
+ * each linear alignment in a chimeric alignment, the read will be hard |
+ * clipped. The alignedSequence and alignedQuality fields in the alignment |
+ * record will only represent the bases for its respective linear alignment. |
*/ |
- core.int templateLength; |
+ core.bool supplementaryAlignment; |
Read(); |
Read.fromJson(core.Map _json) { |
- if (_json.containsKey("alignedBases")) { |
- alignedBases = _json["alignedBases"]; |
+ if (_json.containsKey("alignedQuality")) { |
+ alignedQuality = _json["alignedQuality"]; |
} |
- if (_json.containsKey("baseQuality")) { |
- baseQuality = _json["baseQuality"]; |
+ if (_json.containsKey("alignedSequence")) { |
+ alignedSequence = _json["alignedSequence"]; |
} |
- if (_json.containsKey("cigar")) { |
- cigar = _json["cigar"]; |
+ if (_json.containsKey("alignment")) { |
+ alignment = new LinearAlignment.fromJson(_json["alignment"]); |
} |
- if (_json.containsKey("flags")) { |
- flags = _json["flags"]; |
+ if (_json.containsKey("duplicateFragment")) { |
+ duplicateFragment = _json["duplicateFragment"]; |
+ } |
+ if (_json.containsKey("failedVendorQualityChecks")) { |
+ failedVendorQualityChecks = _json["failedVendorQualityChecks"]; |
+ } |
+ if (_json.containsKey("fragmentLength")) { |
+ fragmentLength = _json["fragmentLength"]; |
+ } |
+ if (_json.containsKey("fragmentName")) { |
+ fragmentName = _json["fragmentName"]; |
} |
if (_json.containsKey("id")) { |
id = _json["id"]; |
} |
- if (_json.containsKey("mappingQuality")) { |
- mappingQuality = _json["mappingQuality"]; |
- } |
- if (_json.containsKey("matePosition")) { |
- matePosition = _json["matePosition"]; |
+ if (_json.containsKey("info")) { |
+ info = _json["info"]; |
} |
- if (_json.containsKey("mateReferenceSequenceName")) { |
- mateReferenceSequenceName = _json["mateReferenceSequenceName"]; |
+ if (_json.containsKey("nextMatePosition")) { |
+ nextMatePosition = new Position.fromJson(_json["nextMatePosition"]); |
} |
- if (_json.containsKey("name")) { |
- name = _json["name"]; |
+ if (_json.containsKey("numberReads")) { |
+ numberReads = _json["numberReads"]; |
} |
- if (_json.containsKey("originalBases")) { |
- originalBases = _json["originalBases"]; |
+ if (_json.containsKey("properPlacement")) { |
+ properPlacement = _json["properPlacement"]; |
} |
- if (_json.containsKey("position")) { |
- position = _json["position"]; |
+ if (_json.containsKey("readGroupId")) { |
+ readGroupId = _json["readGroupId"]; |
} |
- if (_json.containsKey("readsetId")) { |
- readsetId = _json["readsetId"]; |
+ if (_json.containsKey("readGroupSetId")) { |
+ readGroupSetId = _json["readGroupSetId"]; |
} |
- if (_json.containsKey("referenceSequenceName")) { |
- referenceSequenceName = _json["referenceSequenceName"]; |
+ if (_json.containsKey("readNumber")) { |
+ readNumber = _json["readNumber"]; |
} |
- if (_json.containsKey("tags")) { |
- tags = _json["tags"]; |
+ if (_json.containsKey("secondaryAlignment")) { |
+ secondaryAlignment = _json["secondaryAlignment"]; |
} |
- if (_json.containsKey("templateLength")) { |
- templateLength = _json["templateLength"]; |
+ if (_json.containsKey("supplementaryAlignment")) { |
+ supplementaryAlignment = _json["supplementaryAlignment"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (alignedBases != null) { |
- _json["alignedBases"] = alignedBases; |
+ if (alignedQuality != null) { |
+ _json["alignedQuality"] = alignedQuality; |
} |
- if (baseQuality != null) { |
- _json["baseQuality"] = baseQuality; |
+ if (alignedSequence != null) { |
+ _json["alignedSequence"] = alignedSequence; |
} |
- if (cigar != null) { |
- _json["cigar"] = cigar; |
+ if (alignment != null) { |
+ _json["alignment"] = (alignment).toJson(); |
+ } |
+ if (duplicateFragment != null) { |
+ _json["duplicateFragment"] = duplicateFragment; |
+ } |
+ if (failedVendorQualityChecks != null) { |
+ _json["failedVendorQualityChecks"] = failedVendorQualityChecks; |
+ } |
+ if (fragmentLength != null) { |
+ _json["fragmentLength"] = fragmentLength; |
} |
- if (flags != null) { |
- _json["flags"] = flags; |
+ if (fragmentName != null) { |
+ _json["fragmentName"] = fragmentName; |
} |
if (id != null) { |
_json["id"] = id; |
} |
- if (mappingQuality != null) { |
- _json["mappingQuality"] = mappingQuality; |
- } |
- if (matePosition != null) { |
- _json["matePosition"] = matePosition; |
+ if (info != null) { |
+ _json["info"] = info; |
} |
- if (mateReferenceSequenceName != null) { |
- _json["mateReferenceSequenceName"] = mateReferenceSequenceName; |
+ if (nextMatePosition != null) { |
+ _json["nextMatePosition"] = (nextMatePosition).toJson(); |
} |
- if (name != null) { |
- _json["name"] = name; |
+ if (numberReads != null) { |
+ _json["numberReads"] = numberReads; |
} |
- if (originalBases != null) { |
- _json["originalBases"] = originalBases; |
+ if (properPlacement != null) { |
+ _json["properPlacement"] = properPlacement; |
} |
- if (position != null) { |
- _json["position"] = position; |
+ if (readGroupId != null) { |
+ _json["readGroupId"] = readGroupId; |
} |
- if (readsetId != null) { |
- _json["readsetId"] = readsetId; |
+ if (readGroupSetId != null) { |
+ _json["readGroupSetId"] = readGroupSetId; |
} |
- if (referenceSequenceName != null) { |
- _json["referenceSequenceName"] = referenceSequenceName; |
+ if (readNumber != null) { |
+ _json["readNumber"] = readNumber; |
} |
- if (tags != null) { |
- _json["tags"] = tags; |
+ if (secondaryAlignment != null) { |
+ _json["secondaryAlignment"] = secondaryAlignment; |
} |
- if (templateLength != null) { |
- _json["templateLength"] = templateLength; |
+ if (supplementaryAlignment != null) { |
+ _json["supplementaryAlignment"] = supplementaryAlignment; |
} |
return _json; |
} |
} |
-/** Not documented yet. */ |
+/** |
+ * A read group is all the data that's processed the same way by the sequencer. |
+ */ |
class ReadGroup { |
- /** (DT) Date the run was produced (ISO8601 date or date/time). */ |
- core.String date; |
+ /** The ID of the dataset this read group belongs to. */ |
+ core.String datasetId; |
- /** (DS) Description. */ |
+ /** A free-form text description of this read group. */ |
core.String description; |
+ /** The experiment used to generate this read group. */ |
+ ReadGroupExperiment experiment; |
+ |
/** |
- * (FO) Flow order. The array of nucleotide bases that correspond to the |
- * nucleotides used for each flow of each read. |
+ * The generated unique read group ID. Note: This is different than the @RG ID |
+ * field in the SAM spec. For that value, see the name field. |
*/ |
- core.String flowOrder; |
- |
- /** (ID) Read group identifier. */ |
core.String id; |
+ /** A map of additional read group information. */ |
+ core.Map<core.String, core.List<core.String>> info; |
+ |
/** |
- * (KS) The array of nucleotide bases that correspond to the key sequence of |
- * each read. |
+ * The read group name. This corresponds to the @RG ID field in the SAM spec. |
*/ |
- core.String keySequence; |
- |
- /** (LS) Library. */ |
- core.String library; |
- |
- /** (PU) Platform unit. */ |
- core.String platformUnit; |
+ core.String name; |
- /** (PI) Predicted median insert size. */ |
+ /** |
+ * The predicted insert size of this read group. The insert size is the length |
+ * the sequenced DNA fragment from end-to-end, not including the adapters. |
+ */ |
core.int predictedInsertSize; |
- /** (PG) Programs used for processing the read group. */ |
- core.String processingProgram; |
- |
- /** (SM) Sample. */ |
- core.String sample; |
+ /** |
+ * The programs used to generate this read group. Programs are always |
+ * identical for all read groups within a read group set. For this reason, |
+ * only the first read group in a returned set will have this field populated. |
+ */ |
+ core.List<ReadGroupProgram> programs; |
- /** (CN) Name of sequencing center producing the read. */ |
- core.String sequencingCenterName; |
+ /** |
+ * The reference set the reads in this read group are aligned to. Required if |
+ * there are any read alignments. |
+ */ |
+ core.String referenceSetId; |
- /** (PL) Platform/technology used to produce the reads. */ |
- core.String sequencingTechnology; |
+ /** |
+ * The sample this read group's data was generated from. Note: This is not an |
+ * actual ID within this repository, but rather an identifier for a sample |
+ * which may be meaningful to some external system. |
+ */ |
+ core.String sampleId; |
ReadGroup(); |
ReadGroup.fromJson(core.Map _json) { |
- if (_json.containsKey("date")) { |
- date = _json["date"]; |
+ if (_json.containsKey("datasetId")) { |
+ datasetId = _json["datasetId"]; |
} |
if (_json.containsKey("description")) { |
description = _json["description"]; |
} |
- if (_json.containsKey("flowOrder")) { |
- flowOrder = _json["flowOrder"]; |
+ if (_json.containsKey("experiment")) { |
+ experiment = new ReadGroupExperiment.fromJson(_json["experiment"]); |
} |
if (_json.containsKey("id")) { |
id = _json["id"]; |
} |
- if (_json.containsKey("keySequence")) { |
- keySequence = _json["keySequence"]; |
- } |
- if (_json.containsKey("library")) { |
- library = _json["library"]; |
+ if (_json.containsKey("info")) { |
+ info = _json["info"]; |
} |
- if (_json.containsKey("platformUnit")) { |
- platformUnit = _json["platformUnit"]; |
+ if (_json.containsKey("name")) { |
+ name = _json["name"]; |
} |
if (_json.containsKey("predictedInsertSize")) { |
predictedInsertSize = _json["predictedInsertSize"]; |
} |
- if (_json.containsKey("processingProgram")) { |
- processingProgram = _json["processingProgram"]; |
- } |
- if (_json.containsKey("sample")) { |
- sample = _json["sample"]; |
+ if (_json.containsKey("programs")) { |
+ programs = _json["programs"].map((value) => new ReadGroupProgram.fromJson(value)).toList(); |
} |
- if (_json.containsKey("sequencingCenterName")) { |
- sequencingCenterName = _json["sequencingCenterName"]; |
+ if (_json.containsKey("referenceSetId")) { |
+ referenceSetId = _json["referenceSetId"]; |
} |
- if (_json.containsKey("sequencingTechnology")) { |
- sequencingTechnology = _json["sequencingTechnology"]; |
+ if (_json.containsKey("sampleId")) { |
+ sampleId = _json["sampleId"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (date != null) { |
- _json["date"] = date; |
+ if (datasetId != null) { |
+ _json["datasetId"] = datasetId; |
} |
if (description != null) { |
_json["description"] = description; |
} |
- if (flowOrder != null) { |
- _json["flowOrder"] = flowOrder; |
+ if (experiment != null) { |
+ _json["experiment"] = (experiment).toJson(); |
} |
if (id != null) { |
_json["id"] = id; |
} |
- if (keySequence != null) { |
- _json["keySequence"] = keySequence; |
+ if (info != null) { |
+ _json["info"] = info; |
+ } |
+ if (name != null) { |
+ _json["name"] = name; |
+ } |
+ if (predictedInsertSize != null) { |
+ _json["predictedInsertSize"] = predictedInsertSize; |
+ } |
+ if (programs != null) { |
+ _json["programs"] = programs.map((value) => (value).toJson()).toList(); |
+ } |
+ if (referenceSetId != null) { |
+ _json["referenceSetId"] = referenceSetId; |
+ } |
+ if (sampleId != null) { |
+ _json["sampleId"] = sampleId; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ReadGroupExperiment { |
+ /** |
+ * The instrument model used as part of this experiment. This maps to |
+ * sequencing technology in BAM. |
+ */ |
+ core.String instrumentModel; |
+ |
+ /** |
+ * The library used as part of this experiment. Note: This is not an actual ID |
+ * within this repository, but rather an identifier for a library which may be |
+ * meaningful to some external system. |
+ */ |
+ core.String libraryId; |
+ |
+ /** |
+ * The platform unit used as part of this experiment e.g. |
+ * flowcell-barcode.lane for Illumina or slide for SOLiD. Corresponds to the |
+ */ |
+ core.String platformUnit; |
+ |
+ /** The sequencing center used as part of this experiment. */ |
+ core.String sequencingCenter; |
+ |
+ |
+ ReadGroupExperiment(); |
+ |
+ ReadGroupExperiment.fromJson(core.Map _json) { |
+ if (_json.containsKey("instrumentModel")) { |
+ instrumentModel = _json["instrumentModel"]; |
+ } |
+ if (_json.containsKey("libraryId")) { |
+ libraryId = _json["libraryId"]; |
+ } |
+ if (_json.containsKey("platformUnit")) { |
+ platformUnit = _json["platformUnit"]; |
+ } |
+ if (_json.containsKey("sequencingCenter")) { |
+ sequencingCenter = _json["sequencingCenter"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (instrumentModel != null) { |
+ _json["instrumentModel"] = instrumentModel; |
} |
- if (library != null) { |
- _json["library"] = library; |
+ if (libraryId != null) { |
+ _json["libraryId"] = libraryId; |
} |
if (platformUnit != null) { |
_json["platformUnit"] = platformUnit; |
} |
- if (predictedInsertSize != null) { |
- _json["predictedInsertSize"] = predictedInsertSize; |
+ if (sequencingCenter != null) { |
+ _json["sequencingCenter"] = sequencingCenter; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class ReadGroupProgram { |
+ /** The command line used to run this program. */ |
+ core.String commandLine; |
+ |
+ /** |
+ * The user specified locally unique ID of the program. Used along with |
+ * prevProgramId to define an ordering between programs. |
+ */ |
+ core.String id; |
+ |
+ /** The name of the program. */ |
+ core.String name; |
+ |
+ /** The ID of the program run before this one. */ |
+ core.String prevProgramId; |
+ |
+ /** The version of the program run. */ |
+ core.String version; |
+ |
+ |
+ ReadGroupProgram(); |
+ |
+ ReadGroupProgram.fromJson(core.Map _json) { |
+ if (_json.containsKey("commandLine")) { |
+ commandLine = _json["commandLine"]; |
} |
- if (processingProgram != null) { |
- _json["processingProgram"] = processingProgram; |
+ if (_json.containsKey("id")) { |
+ id = _json["id"]; |
+ } |
+ if (_json.containsKey("name")) { |
+ name = _json["name"]; |
+ } |
+ if (_json.containsKey("prevProgramId")) { |
+ prevProgramId = _json["prevProgramId"]; |
+ } |
+ if (_json.containsKey("version")) { |
+ version = _json["version"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (commandLine != null) { |
+ _json["commandLine"] = commandLine; |
+ } |
+ if (id != null) { |
+ _json["id"] = id; |
} |
- if (sample != null) { |
- _json["sample"] = sample; |
+ if (name != null) { |
+ _json["name"] = name; |
} |
- if (sequencingCenterName != null) { |
- _json["sequencingCenterName"] = sequencingCenterName; |
+ if (prevProgramId != null) { |
+ _json["prevProgramId"] = prevProgramId; |
} |
- if (sequencingTechnology != null) { |
- _json["sequencingTechnology"] = sequencingTechnology; |
+ if (version != null) { |
+ _json["version"] = version; |
} |
return _json; |
} |
} |
-/** A Readset is a collection of Reads. */ |
-class Readset { |
- /** The ID of the dataset this readset belongs to. */ |
+/** |
+ * A read group set is a logical collection of read groups, which are |
+ * collections of reads produced by a sequencer. A read group set typically |
+ * models reads corresponding to one sample, sequenced one way, and aligned one |
+ * way. |
+ * - A read group set belongs to one dataset. |
+ * - A read group belongs to one read group set. |
+ * - A read belongs to one read group. |
+ */ |
+class ReadGroupSet { |
+ /** The dataset ID. */ |
core.String datasetId; |
/** |
- * File information from the original BAM import. See the BAM format |
- * specification for additional information on each field. |
+ * The filename of the original source file for this read group set, if any. |
*/ |
- core.List<HeaderSection> fileData; |
+ core.String filename; |
- /** The Google generated ID of the readset, immutable. */ |
+ /** The read group set ID. */ |
core.String id; |
- /** The readset name, typically this is the sample name. */ |
+ /** |
+ * The read group set name. By default this will be initialized to the sample |
+ * name of the sequenced data contained in this set. |
+ */ |
core.String name; |
+ /** |
+ * The read groups in this set. There are typically 1-10 read groups in a read |
+ * group set. |
+ */ |
+ core.List<ReadGroup> readGroups; |
+ |
+ /** The reference set the reads in this read group set are aligned to. */ |
+ core.String referenceSetId; |
- Readset(); |
- Readset.fromJson(core.Map _json) { |
+ ReadGroupSet(); |
+ |
+ ReadGroupSet.fromJson(core.Map _json) { |
if (_json.containsKey("datasetId")) { |
datasetId = _json["datasetId"]; |
} |
- if (_json.containsKey("fileData")) { |
- fileData = _json["fileData"].map((value) => new HeaderSection.fromJson(value)).toList(); |
+ if (_json.containsKey("filename")) { |
+ filename = _json["filename"]; |
} |
if (_json.containsKey("id")) { |
id = _json["id"]; |
@@ -3608,6 +4461,12 @@ class Readset { |
if (_json.containsKey("name")) { |
name = _json["name"]; |
} |
+ if (_json.containsKey("readGroups")) { |
+ readGroups = _json["readGroups"].map((value) => new ReadGroup.fromJson(value)).toList(); |
+ } |
+ if (_json.containsKey("referenceSetId")) { |
+ referenceSetId = _json["referenceSetId"]; |
+ } |
} |
core.Map toJson() { |
@@ -3615,15 +4474,117 @@ class Readset { |
if (datasetId != null) { |
_json["datasetId"] = datasetId; |
} |
- if (fileData != null) { |
- _json["fileData"] = fileData.map((value) => (value).toJson()).toList(); |
+ if (filename != null) { |
+ _json["filename"] = filename; |
+ } |
+ if (id != null) { |
+ _json["id"] = id; |
+ } |
+ if (name != null) { |
+ _json["name"] = name; |
+ } |
+ if (readGroups != null) { |
+ _json["readGroups"] = readGroups.map((value) => (value).toJson()).toList(); |
+ } |
+ if (referenceSetId != null) { |
+ _json["referenceSetId"] = referenceSetId; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** |
+ * A reference is a canonical assembled DNA sequence, intended to act as a |
+ * reference coordinate space for other genomic annotations. A single reference |
+ * might represent the human chromosome 1 or mitochandrial DNA, for instance. A |
+ * reference belongs to one or more reference sets. |
+ */ |
+class Reference { |
+ /** The Google generated immutable ID of the reference. */ |
+ core.String id; |
+ |
+ /** The length of this reference's sequence. */ |
+ core.String length; |
+ |
+ /** |
+ * MD5 of the upper-case sequence excluding all whitespace characters (this is |
+ * equivalent to SQ:M5 in SAM). This value is represented in lower case |
+ * hexadecimal format. |
+ */ |
+ core.String md5checksum; |
+ |
+ /** The name of this reference, for example 22. */ |
+ core.String name; |
+ |
+ /** |
+ * ID from http://www.ncbi.nlm.nih.gov/taxonomy (e.g. 9606->human) if not |
+ * specified by the containing reference set. |
+ */ |
+ core.int ncbiTaxonId; |
+ |
+ /** |
+ * All known corresponding accession IDs in INSDC (GenBank/ENA/DDBJ) ideally |
+ * with a version number, for example GCF_000001405.26. |
+ */ |
+ core.List<core.String> sourceAccessions; |
+ |
+ /** |
+ * The URI from which the sequence was obtained. Specifies a FASTA format |
+ * file/string with one name, sequence pair. |
+ */ |
+ core.String sourceURI; |
+ |
+ |
+ Reference(); |
+ |
+ Reference.fromJson(core.Map _json) { |
+ if (_json.containsKey("id")) { |
+ id = _json["id"]; |
+ } |
+ if (_json.containsKey("length")) { |
+ length = _json["length"]; |
+ } |
+ if (_json.containsKey("md5checksum")) { |
+ md5checksum = _json["md5checksum"]; |
} |
+ if (_json.containsKey("name")) { |
+ name = _json["name"]; |
+ } |
+ if (_json.containsKey("ncbiTaxonId")) { |
+ ncbiTaxonId = _json["ncbiTaxonId"]; |
+ } |
+ if (_json.containsKey("sourceAccessions")) { |
+ sourceAccessions = _json["sourceAccessions"]; |
+ } |
+ if (_json.containsKey("sourceURI")) { |
+ sourceURI = _json["sourceURI"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
if (id != null) { |
_json["id"] = id; |
} |
+ if (length != null) { |
+ _json["length"] = length; |
+ } |
+ if (md5checksum != null) { |
+ _json["md5checksum"] = md5checksum; |
+ } |
if (name != null) { |
_json["name"] = name; |
} |
+ if (ncbiTaxonId != null) { |
+ _json["ncbiTaxonId"] = ncbiTaxonId; |
+ } |
+ if (sourceAccessions != null) { |
+ _json["sourceAccessions"] = sourceAccessions; |
+ } |
+ if (sourceURI != null) { |
+ _json["sourceURI"] = sourceURI; |
+ } |
return _json; |
} |
} |
@@ -3668,50 +4629,83 @@ class ReferenceBound { |
} |
-/** Not documented yet. */ |
-class ReferenceSequence { |
- /** (AS) Genome assembly identifier. */ |
+/** |
+ * A reference set is a set of references which typically comprise a reference |
+ * assembly for a species, such as GRCh38 which is representative of the human |
+ * genome. A reference set defines a common coordinate space for comparing |
+ * reference-aligned experimental data. A reference set contains 1 or more |
+ * references. |
+ */ |
+class ReferenceSet { |
+ /** Public id of this reference set, such as GRCh37. */ |
core.String assemblyId; |
- /** (LN) Reference sequence length. */ |
- core.int length; |
+ /** Optional free text description of this reference set. */ |
+ core.String description; |
+ |
+ /** The Google generated immutable ID of the reference set. */ |
+ core.String id; |
/** |
- * (M5) MD5 checksum of the sequence in the uppercase, excluding spaces but |
- * including pads as *. |
+ * Order-independent MD5 checksum which identifies this reference set. The |
+ * checksum is computed by sorting all lower case hexidecimal string |
+ * reference.md5checksum (for all reference in this set) in ascending |
+ * lexicographic order, concatenating, and taking the MD5 of that value. The |
+ * resulting value is represented in lower case hexadecimal format. |
*/ |
- core.String md5Checksum; |
+ core.String md5checksum; |
- /** (SN) Reference sequence name. */ |
- core.String name; |
+ /** |
+ * ID from http://www.ncbi.nlm.nih.gov/taxonomy (e.g. 9606->human) indicating |
+ * the species which this assembly is intended to model. Note that contained |
+ * references may specify a different ncbiTaxonId, as assemblies may contain |
+ * reference sequences which do not belong to the modeled species, e.g. EBV in |
+ * a human reference genome. |
+ */ |
+ core.int ncbiTaxonId; |
+ |
+ /** |
+ * The IDs of the reference objects that are part of this set. |
+ * Reference.md5checksum must be unique within this set. |
+ */ |
+ core.List<core.String> referenceIds; |
- /** (SP) Species. */ |
- core.String species; |
+ /** |
+ * All known corresponding accession IDs in INSDC (GenBank/ENA/DDBJ) ideally |
+ * with a version number, for example NC_000001.11. |
+ */ |
+ core.List<core.String> sourceAccessions; |
- /** (UR) URI of the sequence. */ |
- core.String uri; |
+ /** The URI from which the references were obtained. */ |
+ core.String sourceURI; |
- ReferenceSequence(); |
+ ReferenceSet(); |
- ReferenceSequence.fromJson(core.Map _json) { |
+ ReferenceSet.fromJson(core.Map _json) { |
if (_json.containsKey("assemblyId")) { |
assemblyId = _json["assemblyId"]; |
} |
- if (_json.containsKey("length")) { |
- length = _json["length"]; |
+ if (_json.containsKey("description")) { |
+ description = _json["description"]; |
+ } |
+ if (_json.containsKey("id")) { |
+ id = _json["id"]; |
} |
- if (_json.containsKey("md5Checksum")) { |
- md5Checksum = _json["md5Checksum"]; |
+ if (_json.containsKey("md5checksum")) { |
+ md5checksum = _json["md5checksum"]; |
} |
- if (_json.containsKey("name")) { |
- name = _json["name"]; |
+ if (_json.containsKey("ncbiTaxonId")) { |
+ ncbiTaxonId = _json["ncbiTaxonId"]; |
} |
- if (_json.containsKey("species")) { |
- species = _json["species"]; |
+ if (_json.containsKey("referenceIds")) { |
+ referenceIds = _json["referenceIds"]; |
} |
- if (_json.containsKey("uri")) { |
- uri = _json["uri"]; |
+ if (_json.containsKey("sourceAccessions")) { |
+ sourceAccessions = _json["sourceAccessions"]; |
+ } |
+ if (_json.containsKey("sourceURI")) { |
+ sourceURI = _json["sourceURI"]; |
} |
} |
@@ -3720,20 +4714,26 @@ class ReferenceSequence { |
if (assemblyId != null) { |
_json["assemblyId"] = assemblyId; |
} |
- if (length != null) { |
- _json["length"] = length; |
+ if (description != null) { |
+ _json["description"] = description; |
} |
- if (md5Checksum != null) { |
- _json["md5Checksum"] = md5Checksum; |
+ if (id != null) { |
+ _json["id"] = id; |
} |
- if (name != null) { |
- _json["name"] = name; |
+ if (md5checksum != null) { |
+ _json["md5checksum"] = md5checksum; |
+ } |
+ if (ncbiTaxonId != null) { |
+ _json["ncbiTaxonId"] = ncbiTaxonId; |
} |
- if (species != null) { |
- _json["species"] = species; |
+ if (referenceIds != null) { |
+ _json["referenceIds"] = referenceIds; |
} |
- if (uri != null) { |
- _json["uri"] = uri; |
+ if (sourceAccessions != null) { |
+ _json["sourceAccessions"] = sourceAccessions; |
+ } |
+ if (sourceURI != null) { |
+ _json["sourceURI"] = sourceURI; |
} |
return _json; |
} |
@@ -3856,7 +4856,7 @@ class SearchJobsRequest { |
* Specifies the number of results to return in a single page. Defaults to |
* 128. The maximum value is 256. |
*/ |
- core.String maxResults; |
+ core.int pageSize; |
/** |
* The continuation token which is used to page through large result sets. To |
@@ -3866,7 +4866,7 @@ class SearchJobsRequest { |
core.String pageToken; |
/** Required. Only return jobs which belong to this Google Developers */ |
- core.String projectId; |
+ core.String projectNumber; |
/** Only return jobs which have a matching status. */ |
core.List<core.String> status; |
@@ -3881,14 +4881,14 @@ class SearchJobsRequest { |
if (_json.containsKey("createdBefore")) { |
createdBefore = _json["createdBefore"]; |
} |
- if (_json.containsKey("maxResults")) { |
- maxResults = _json["maxResults"]; |
+ if (_json.containsKey("pageSize")) { |
+ pageSize = _json["pageSize"]; |
} |
if (_json.containsKey("pageToken")) { |
pageToken = _json["pageToken"]; |
} |
- if (_json.containsKey("projectId")) { |
- projectId = _json["projectId"]; |
+ if (_json.containsKey("projectNumber")) { |
+ projectNumber = _json["projectNumber"]; |
} |
if (_json.containsKey("status")) { |
status = _json["status"]; |
@@ -3903,14 +4903,14 @@ class SearchJobsRequest { |
if (createdBefore != null) { |
_json["createdBefore"] = createdBefore; |
} |
- if (maxResults != null) { |
- _json["maxResults"] = maxResults; |
+ if (pageSize != null) { |
+ _json["pageSize"] = pageSize; |
} |
if (pageToken != null) { |
_json["pageToken"] = pageToken; |
} |
- if (projectId != null) { |
- _json["projectId"] = projectId; |
+ if (projectNumber != null) { |
+ _json["projectNumber"] = projectNumber; |
} |
if (status != null) { |
_json["status"] = status; |
@@ -3957,13 +4957,120 @@ class SearchJobsResponse { |
} |
+/** The read group set search request. */ |
+class SearchReadGroupSetsRequest { |
+ /** |
+ * Restricts this query to read group sets within the given datasets. At least |
+ * one ID must be provided. |
+ */ |
+ core.List<core.String> datasetIds; |
+ |
+ /** |
+ * Only return read group sets for which a substring of the name matches this |
+ * string. |
+ */ |
+ core.String name; |
+ |
+ /** |
+ * Specifies number of results to return in a single page. If unspecified, it |
+ * will default to 128. The maximum value is 1024. |
+ */ |
+ core.int pageSize; |
+ |
+ /** |
+ * The continuation token, which is used to page through large result sets. To |
+ * get the next page of results, set this parameter to the value of |
+ * nextPageToken from the previous response. |
+ */ |
+ core.String pageToken; |
+ |
+ |
+ SearchReadGroupSetsRequest(); |
+ |
+ SearchReadGroupSetsRequest.fromJson(core.Map _json) { |
+ if (_json.containsKey("datasetIds")) { |
+ datasetIds = _json["datasetIds"]; |
+ } |
+ if (_json.containsKey("name")) { |
+ name = _json["name"]; |
+ } |
+ if (_json.containsKey("pageSize")) { |
+ pageSize = _json["pageSize"]; |
+ } |
+ if (_json.containsKey("pageToken")) { |
+ pageToken = _json["pageToken"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (datasetIds != null) { |
+ _json["datasetIds"] = datasetIds; |
+ } |
+ if (name != null) { |
+ _json["name"] = name; |
+ } |
+ if (pageSize != null) { |
+ _json["pageSize"] = pageSize; |
+ } |
+ if (pageToken != null) { |
+ _json["pageToken"] = pageToken; |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** The read group set search response. */ |
+class SearchReadGroupSetsResponse { |
+ /** |
+ * The continuation token, which is used to page through large result sets. |
+ * Provide this value in a subsequent request to return the next page of |
+ * results. This field will be empty if there aren't any additional results. |
+ */ |
+ core.String nextPageToken; |
+ |
+ /** The list of matching read group sets. */ |
+ core.List<ReadGroupSet> readGroupSets; |
+ |
+ |
+ SearchReadGroupSetsResponse(); |
+ |
+ SearchReadGroupSetsResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("nextPageToken")) { |
+ nextPageToken = _json["nextPageToken"]; |
+ } |
+ if (_json.containsKey("readGroupSets")) { |
+ readGroupSets = _json["readGroupSets"].map((value) => new ReadGroupSet.fromJson(value)).toList(); |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (nextPageToken != null) { |
+ _json["nextPageToken"] = nextPageToken; |
+ } |
+ if (readGroupSets != null) { |
+ _json["readGroupSets"] = readGroupSets.map((value) => (value).toJson()).toList(); |
+ } |
+ return _json; |
+ } |
+} |
+ |
+ |
/** The read search request. */ |
class SearchReadsRequest { |
/** |
+ * The end position of the range on the reference, 0-based exclusive. If |
+ * specified, referenceName must also be specified. |
+ */ |
+ core.String end; |
+ |
+ /** |
* Specifies number of results to return in a single page. If unspecified, it |
* will default to 256. The maximum value is 2048. |
*/ |
- core.String maxResults; |
+ core.int pageSize; |
/** |
* The continuation token, which is used to page through large result sets. To |
@@ -3973,76 +5080,81 @@ class SearchReadsRequest { |
core.String pageToken; |
/** |
- * The readsets within which to search for reads. At least one readset ID must |
- * be provided. All specified readsets must be aligned against a common set of |
- * reference sequences; this defines the genomic coordinates for the query. |
+ * The IDs of the read groups within which to search for reads. All specified |
+ * read groups must belong to the same read group sets. Must specify one of |
+ * readGroupSetIds or readGroupIds. |
*/ |
- core.List<core.String> readsetIds; |
+ core.List<core.String> readGroupIds; |
/** |
- * The end position (1-based, inclusive) of the target range. If specified, |
- * sequenceName must also be specified. Defaults to the end of the target |
- * reference sequence, if any. |
+ * The IDs of the read groups sets within which to search for reads. All |
+ * specified read group sets must be aligned against a common set of reference |
+ * sequences; this defines the genomic coordinates for the query. Must specify |
+ * one of readGroupSetIds or readGroupIds. |
*/ |
- core.String sequenceEnd; |
+ core.List<core.String> readGroupSetIds; |
/** |
- * Restricts the results to a particular reference sequence such as 1, chr1, |
- * or X. The set of valid references sequences depends on the readsets |
- * specified. If set to *, only unmapped Reads are returned. |
+ * The reference sequence name, for example chr1, 1, or chrX. If set to *, |
+ * only unmapped reads are returned. |
*/ |
- core.String sequenceName; |
+ core.String referenceName; |
/** |
- * The start position (1-based, inclusive) of the target range. If specified, |
- * sequenceName must also be specified. Defaults to the start of the target |
- * reference sequence, if any. |
+ * The start position of the range on the reference, 0-based inclusive. If |
+ * specified, referenceName must also be specified. |
*/ |
- core.String sequenceStart; |
+ core.String start; |
SearchReadsRequest(); |
SearchReadsRequest.fromJson(core.Map _json) { |
- if (_json.containsKey("maxResults")) { |
- maxResults = _json["maxResults"]; |
+ if (_json.containsKey("end")) { |
+ end = _json["end"]; |
+ } |
+ if (_json.containsKey("pageSize")) { |
+ pageSize = _json["pageSize"]; |
} |
if (_json.containsKey("pageToken")) { |
pageToken = _json["pageToken"]; |
} |
- if (_json.containsKey("readsetIds")) { |
- readsetIds = _json["readsetIds"]; |
+ if (_json.containsKey("readGroupIds")) { |
+ readGroupIds = _json["readGroupIds"]; |
} |
- if (_json.containsKey("sequenceEnd")) { |
- sequenceEnd = _json["sequenceEnd"]; |
+ if (_json.containsKey("readGroupSetIds")) { |
+ readGroupSetIds = _json["readGroupSetIds"]; |
} |
- if (_json.containsKey("sequenceName")) { |
- sequenceName = _json["sequenceName"]; |
+ if (_json.containsKey("referenceName")) { |
+ referenceName = _json["referenceName"]; |
} |
- if (_json.containsKey("sequenceStart")) { |
- sequenceStart = _json["sequenceStart"]; |
+ if (_json.containsKey("start")) { |
+ start = _json["start"]; |
} |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (maxResults != null) { |
- _json["maxResults"] = maxResults; |
+ if (end != null) { |
+ _json["end"] = end; |
+ } |
+ if (pageSize != null) { |
+ _json["pageSize"] = pageSize; |
} |
if (pageToken != null) { |
_json["pageToken"] = pageToken; |
} |
- if (readsetIds != null) { |
- _json["readsetIds"] = readsetIds; |
+ if (readGroupIds != null) { |
+ _json["readGroupIds"] = readGroupIds; |
} |
- if (sequenceEnd != null) { |
- _json["sequenceEnd"] = sequenceEnd; |
+ if (readGroupSetIds != null) { |
+ _json["readGroupSetIds"] = readGroupSetIds; |
} |
- if (sequenceName != null) { |
- _json["sequenceName"] = sequenceName; |
+ if (referenceName != null) { |
+ _json["referenceName"] = referenceName; |
} |
- if (sequenceStart != null) { |
- _json["sequenceStart"] = sequenceStart; |
+ if (start != null) { |
+ _json["start"] = start; |
} |
return _json; |
} |
@@ -4052,62 +5164,165 @@ class SearchReadsRequest { |
/** The read search response. */ |
class SearchReadsResponse { |
/** |
+ * The list of matching alignments sorted by mapped genomic coordinate, if |
+ * any, ascending in position within the same reference. Unmapped reads, which |
+ * have no position, are returned last and are further sorted in ascending |
+ * lexicographic order by fragment name. |
+ */ |
+ core.List<Read> alignments; |
+ |
+ /** |
* The continuation token, which is used to page through large result sets. |
* Provide this value in a subsequent request to return the next page of |
* results. This field will be empty if there aren't any additional results. |
*/ |
core.String nextPageToken; |
- /** |
- * The list of matching Reads. The resulting Reads are sorted by position; the |
- * smallest positions are returned first. Unmapped reads, which have no |
- * position, are returned last and are further sorted alphabetically by name. |
- */ |
- core.List<Read> reads; |
- |
SearchReadsResponse(); |
SearchReadsResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("alignments")) { |
+ alignments = _json["alignments"].map((value) => new Read.fromJson(value)).toList(); |
+ } |
if (_json.containsKey("nextPageToken")) { |
nextPageToken = _json["nextPageToken"]; |
} |
- if (_json.containsKey("reads")) { |
- reads = _json["reads"].map((value) => new Read.fromJson(value)).toList(); |
- } |
} |
core.Map toJson() { |
var _json = new core.Map(); |
+ if (alignments != null) { |
+ _json["alignments"] = alignments.map((value) => (value).toJson()).toList(); |
+ } |
if (nextPageToken != null) { |
_json["nextPageToken"] = nextPageToken; |
} |
- if (reads != null) { |
- _json["reads"] = reads.map((value) => (value).toJson()).toList(); |
+ return _json; |
+ } |
+} |
+ |
+ |
+/** Not documented yet. */ |
+class SearchReferenceSetsRequest { |
+ /** |
+ * If present, return references for which the accession matches any of these |
+ * strings. Best to give a version number, for example GCF_000001405.26. If |
+ * only the main accession number is given then all records with that main |
+ * accession will be returned, whichever version. Note that different versions |
+ * will have different sequences. |
+ */ |
+ core.List<core.String> accessions; |
+ |
+ /** |
+ * If present, return references for which the md5checksum matches. See |
+ * ReferenceSet.md5checksum for details. |
+ */ |
+ core.List<core.String> md5checksums; |
+ |
+ /** Specifies the maximum number of results to return in a single page. */ |
+ core.int pageSize; |
+ |
+ /** |
+ * The continuation token, which is used to page through large result sets. To |
+ * get the next page of results, set this parameter to the value of |
+ * nextPageToken from the previous response. |
+ */ |
+ core.String pageToken; |
+ |
+ |
+ SearchReferenceSetsRequest(); |
+ |
+ SearchReferenceSetsRequest.fromJson(core.Map _json) { |
+ if (_json.containsKey("accessions")) { |
+ accessions = _json["accessions"]; |
+ } |
+ if (_json.containsKey("md5checksums")) { |
+ md5checksums = _json["md5checksums"]; |
+ } |
+ if (_json.containsKey("pageSize")) { |
+ pageSize = _json["pageSize"]; |
+ } |
+ if (_json.containsKey("pageToken")) { |
+ pageToken = _json["pageToken"]; |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (accessions != null) { |
+ _json["accessions"] = accessions; |
+ } |
+ if (md5checksums != null) { |
+ _json["md5checksums"] = md5checksums; |
+ } |
+ if (pageSize != null) { |
+ _json["pageSize"] = pageSize; |
+ } |
+ if (pageToken != null) { |
+ _json["pageToken"] = pageToken; |
} |
return _json; |
} |
} |
-/** The readset search request. */ |
-class SearchReadsetsRequest { |
+/** Not documented yet. */ |
+class SearchReferenceSetsResponse { |
/** |
- * Restricts this query to readsets within the given datasets. At least one ID |
- * must be provided. |
+ * The continuation token, which is used to page through large result sets. |
+ * Provide this value in a subsequent request to return the next page of |
+ * results. This field will be empty if there aren't any additional results. |
*/ |
- core.List<core.String> datasetIds; |
+ core.String nextPageToken; |
+ |
+ /** The matching references sets. */ |
+ core.List<ReferenceSet> referenceSets; |
+ |
+ |
+ SearchReferenceSetsResponse(); |
+ |
+ SearchReferenceSetsResponse.fromJson(core.Map _json) { |
+ if (_json.containsKey("nextPageToken")) { |
+ nextPageToken = _json["nextPageToken"]; |
+ } |
+ if (_json.containsKey("referenceSets")) { |
+ referenceSets = _json["referenceSets"].map((value) => new ReferenceSet.fromJson(value)).toList(); |
+ } |
+ } |
+ |
+ core.Map toJson() { |
+ var _json = new core.Map(); |
+ if (nextPageToken != null) { |
+ _json["nextPageToken"] = nextPageToken; |
+ } |
+ if (referenceSets != null) { |
+ _json["referenceSets"] = referenceSets.map((value) => (value).toJson()).toList(); |
+ } |
+ return _json; |
+ } |
+} |
+ |
+/** Not documented yet. */ |
+class SearchReferencesRequest { |
/** |
- * Specifies number of results to return in a single page. If unspecified, it |
- * will default to 128. The maximum value is 1024. |
+ * If present, return references for which the accession matches this string. |
+ * Best to give a version number, for example GCF_000001405.26. If only the |
+ * main accession number is given then all records with that main accession |
+ * will be returned, whichever version. Note that different versions will have |
+ * different sequences. |
*/ |
- core.String maxResults; |
+ core.List<core.String> accessions; |
/** |
- * Only return readsets for which a substring of the name matches this string. |
+ * If present, return references for which the md5checksum matches. See |
+ * Reference.md5checksum for construction details. |
*/ |
- core.String name; |
+ core.List<core.String> md5checksums; |
+ |
+ /** Specifies the maximum number of results to return in a single page. */ |
+ core.int pageSize; |
/** |
* The continuation token, which is used to page through large result sets. To |
@@ -4116,45 +5331,54 @@ class SearchReadsetsRequest { |
*/ |
core.String pageToken; |
+ /** If present, return only references which belong to this reference set. */ |
+ core.String referenceSetId; |
- SearchReadsetsRequest(); |
- SearchReadsetsRequest.fromJson(core.Map _json) { |
- if (_json.containsKey("datasetIds")) { |
- datasetIds = _json["datasetIds"]; |
+ SearchReferencesRequest(); |
+ |
+ SearchReferencesRequest.fromJson(core.Map _json) { |
+ if (_json.containsKey("accessions")) { |
+ accessions = _json["accessions"]; |
} |
- if (_json.containsKey("maxResults")) { |
- maxResults = _json["maxResults"]; |
+ if (_json.containsKey("md5checksums")) { |
+ md5checksums = _json["md5checksums"]; |
} |
- if (_json.containsKey("name")) { |
- name = _json["name"]; |
+ if (_json.containsKey("pageSize")) { |
+ pageSize = _json["pageSize"]; |
} |
if (_json.containsKey("pageToken")) { |
pageToken = _json["pageToken"]; |
} |
+ if (_json.containsKey("referenceSetId")) { |
+ referenceSetId = _json["referenceSetId"]; |
+ } |
} |
core.Map toJson() { |
var _json = new core.Map(); |
- if (datasetIds != null) { |
- _json["datasetIds"] = datasetIds; |
+ if (accessions != null) { |
+ _json["accessions"] = accessions; |
} |
- if (maxResults != null) { |
- _json["maxResults"] = maxResults; |
+ if (md5checksums != null) { |
+ _json["md5checksums"] = md5checksums; |
} |
- if (name != null) { |
- _json["name"] = name; |
+ if (pageSize != null) { |
+ _json["pageSize"] = pageSize; |
} |
if (pageToken != null) { |
_json["pageToken"] = pageToken; |
} |
+ if (referenceSetId != null) { |
+ _json["referenceSetId"] = referenceSetId; |
+ } |
return _json; |
} |
} |
-/** The readset search response. */ |
-class SearchReadsetsResponse { |
+/** Not documented yet. */ |
+class SearchReferencesResponse { |
/** |
* The continuation token, which is used to page through large result sets. |
* Provide this value in a subsequent request to return the next page of |
@@ -4162,18 +5386,18 @@ class SearchReadsetsResponse { |
*/ |
core.String nextPageToken; |
- /** The list of matching Readsets. */ |
- core.List<Readset> readsets; |
+ /** The matching references. */ |
+ core.List<Reference> references; |
- SearchReadsetsResponse(); |
+ SearchReferencesResponse(); |
- SearchReadsetsResponse.fromJson(core.Map _json) { |
+ SearchReferencesResponse.fromJson(core.Map _json) { |
if (_json.containsKey("nextPageToken")) { |
nextPageToken = _json["nextPageToken"]; |
} |
- if (_json.containsKey("readsets")) { |
- readsets = _json["readsets"].map((value) => new Readset.fromJson(value)).toList(); |
+ if (_json.containsKey("references")) { |
+ references = _json["references"].map((value) => new Reference.fromJson(value)).toList(); |
} |
} |
@@ -4182,8 +5406,8 @@ class SearchReadsetsResponse { |
if (nextPageToken != null) { |
_json["nextPageToken"] = nextPageToken; |
} |
- if (readsets != null) { |
- _json["readsets"] = readsets.map((value) => (value).toJson()).toList(); |
+ if (references != null) { |
+ _json["references"] = references.map((value) => (value).toJson()).toList(); |
} |
return _json; |
} |
@@ -4431,9 +5655,13 @@ class SearchVariantsResponse { |
/** |
- * A Variant represents a change in DNA sequence relative to some reference. For |
- * example, a Variant could represent a SNP or an insertion. Variants belong to |
- * a variant set. |
+ * A variant represents a change in DNA sequence relative to a reference |
+ * sequence. For example, a variant could represent a SNP or an insertion. |
+ * Variants belong to a variant set. Each of the calls on a variant represent a |
+ * determination of genotype with respect to that variant. For example, a call |
+ * might assign probability of 0.32 to the occurrence of a SNP named rs1234 in a |
+ * sample named NA12345. A call belongs to a call set, which contains related |
+ * calls typically from one sample. |
*/ |
class Variant { |
/** The bases that appear instead of the reference bases. */ |
@@ -4586,8 +5814,8 @@ class Variant { |
/** |
- * A VariantSet represents a collection of Variants and their summary |
- * statistics. |
+ * A variant set is a collection of call sets and variants. It contains summary |
+ * statistics of those contents. A variant set belongs to a dataset. |
*/ |
class VariantSet { |
/** The dataset to which this variant set belongs. Immutable. */ |