| Index: generated/googleapis/lib/bigquery/v2.dart
|
| diff --git a/generated/googleapis/lib/bigquery/v2.dart b/generated/googleapis/lib/bigquery/v2.dart
|
| index 332dcec5cb372cd3f5b46a1a963a2f65bf136f2c..11c817e019d2f2796b7884116034a117f7e37808 100644
|
| --- a/generated/googleapis/lib/bigquery/v2.dart
|
| +++ b/generated/googleapis/lib/bigquery/v2.dart
|
| @@ -2007,6 +2007,10 @@ class ErrorProto {
|
| }
|
|
|
| class ExplainQueryStage {
|
| + /** Milliseconds the average shard spent on CPU-bound tasks. */
|
| + core.String computeMsAvg;
|
| + /** Milliseconds the slowest shard spent on CPU-bound tasks. */
|
| + core.String computeMsMax;
|
| /** Relative amount of time the average shard spent on CPU-bound tasks. */
|
| core.double computeRatioAvg;
|
| /** Relative amount of time the slowest shard spent on CPU-bound tasks. */
|
| @@ -2015,6 +2019,10 @@ class ExplainQueryStage {
|
| core.String id;
|
| /** Human-readable name for stage. */
|
| core.String name;
|
| + /** Milliseconds the average shard spent reading input. */
|
| + core.String readMsAvg;
|
| + /** Milliseconds the slowest shard spent reading input. */
|
| + core.String readMsMax;
|
| /** Relative amount of time the average shard spent reading input. */
|
| core.double readRatioAvg;
|
| /** Relative amount of time the slowest shard spent reading input. */
|
| @@ -2023,6 +2031,10 @@ class ExplainQueryStage {
|
| core.String recordsRead;
|
| /** Number of records written by the stage. */
|
| core.String recordsWritten;
|
| + /** Total number of bytes written to shuffle. */
|
| + core.String shuffleOutputBytes;
|
| + /** Total number of bytes written to shuffle and spilled to disk. */
|
| + core.String shuffleOutputBytesSpilled;
|
| /** Current status for the stage. */
|
| core.String status;
|
| /**
|
| @@ -2030,6 +2042,10 @@ class ExplainQueryStage {
|
| * chronological).
|
| */
|
| core.List<ExplainQueryStep> steps;
|
| + /** Milliseconds the average shard spent waiting to be scheduled. */
|
| + core.String waitMsAvg;
|
| + /** Milliseconds the slowest shard spent waiting to be scheduled. */
|
| + core.String waitMsMax;
|
| /**
|
| * Relative amount of time the average shard spent waiting to be scheduled.
|
| */
|
| @@ -2038,6 +2054,10 @@ class ExplainQueryStage {
|
| * Relative amount of time the slowest shard spent waiting to be scheduled.
|
| */
|
| core.double waitRatioMax;
|
| + /** Milliseconds the average shard spent on writing output. */
|
| + core.String writeMsAvg;
|
| + /** Milliseconds the slowest shard spent on writing output. */
|
| + core.String writeMsMax;
|
| /** Relative amount of time the average shard spent on writing output. */
|
| core.double writeRatioAvg;
|
| /** Relative amount of time the slowest shard spent on writing output. */
|
| @@ -2046,6 +2066,12 @@ class ExplainQueryStage {
|
| ExplainQueryStage();
|
|
|
| ExplainQueryStage.fromJson(core.Map _json) {
|
| + if (_json.containsKey("computeMsAvg")) {
|
| + computeMsAvg = _json["computeMsAvg"];
|
| + }
|
| + if (_json.containsKey("computeMsMax")) {
|
| + computeMsMax = _json["computeMsMax"];
|
| + }
|
| if (_json.containsKey("computeRatioAvg")) {
|
| computeRatioAvg = _json["computeRatioAvg"];
|
| }
|
| @@ -2058,6 +2084,12 @@ class ExplainQueryStage {
|
| if (_json.containsKey("name")) {
|
| name = _json["name"];
|
| }
|
| + if (_json.containsKey("readMsAvg")) {
|
| + readMsAvg = _json["readMsAvg"];
|
| + }
|
| + if (_json.containsKey("readMsMax")) {
|
| + readMsMax = _json["readMsMax"];
|
| + }
|
| if (_json.containsKey("readRatioAvg")) {
|
| readRatioAvg = _json["readRatioAvg"];
|
| }
|
| @@ -2070,18 +2102,36 @@ class ExplainQueryStage {
|
| if (_json.containsKey("recordsWritten")) {
|
| recordsWritten = _json["recordsWritten"];
|
| }
|
| + if (_json.containsKey("shuffleOutputBytes")) {
|
| + shuffleOutputBytes = _json["shuffleOutputBytes"];
|
| + }
|
| + if (_json.containsKey("shuffleOutputBytesSpilled")) {
|
| + shuffleOutputBytesSpilled = _json["shuffleOutputBytesSpilled"];
|
| + }
|
| if (_json.containsKey("status")) {
|
| status = _json["status"];
|
| }
|
| if (_json.containsKey("steps")) {
|
| steps = _json["steps"].map((value) => new ExplainQueryStep.fromJson(value)).toList();
|
| }
|
| + if (_json.containsKey("waitMsAvg")) {
|
| + waitMsAvg = _json["waitMsAvg"];
|
| + }
|
| + if (_json.containsKey("waitMsMax")) {
|
| + waitMsMax = _json["waitMsMax"];
|
| + }
|
| if (_json.containsKey("waitRatioAvg")) {
|
| waitRatioAvg = _json["waitRatioAvg"];
|
| }
|
| if (_json.containsKey("waitRatioMax")) {
|
| waitRatioMax = _json["waitRatioMax"];
|
| }
|
| + if (_json.containsKey("writeMsAvg")) {
|
| + writeMsAvg = _json["writeMsAvg"];
|
| + }
|
| + if (_json.containsKey("writeMsMax")) {
|
| + writeMsMax = _json["writeMsMax"];
|
| + }
|
| if (_json.containsKey("writeRatioAvg")) {
|
| writeRatioAvg = _json["writeRatioAvg"];
|
| }
|
| @@ -2092,6 +2142,12 @@ class ExplainQueryStage {
|
|
|
| core.Map<core.String, core.Object> toJson() {
|
| final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
|
| + if (computeMsAvg != null) {
|
| + _json["computeMsAvg"] = computeMsAvg;
|
| + }
|
| + if (computeMsMax != null) {
|
| + _json["computeMsMax"] = computeMsMax;
|
| + }
|
| if (computeRatioAvg != null) {
|
| _json["computeRatioAvg"] = computeRatioAvg;
|
| }
|
| @@ -2104,6 +2160,12 @@ class ExplainQueryStage {
|
| if (name != null) {
|
| _json["name"] = name;
|
| }
|
| + if (readMsAvg != null) {
|
| + _json["readMsAvg"] = readMsAvg;
|
| + }
|
| + if (readMsMax != null) {
|
| + _json["readMsMax"] = readMsMax;
|
| + }
|
| if (readRatioAvg != null) {
|
| _json["readRatioAvg"] = readRatioAvg;
|
| }
|
| @@ -2116,18 +2178,36 @@ class ExplainQueryStage {
|
| if (recordsWritten != null) {
|
| _json["recordsWritten"] = recordsWritten;
|
| }
|
| + if (shuffleOutputBytes != null) {
|
| + _json["shuffleOutputBytes"] = shuffleOutputBytes;
|
| + }
|
| + if (shuffleOutputBytesSpilled != null) {
|
| + _json["shuffleOutputBytesSpilled"] = shuffleOutputBytesSpilled;
|
| + }
|
| if (status != null) {
|
| _json["status"] = status;
|
| }
|
| if (steps != null) {
|
| _json["steps"] = steps.map((value) => (value).toJson()).toList();
|
| }
|
| + if (waitMsAvg != null) {
|
| + _json["waitMsAvg"] = waitMsAvg;
|
| + }
|
| + if (waitMsMax != null) {
|
| + _json["waitMsMax"] = waitMsMax;
|
| + }
|
| if (waitRatioAvg != null) {
|
| _json["waitRatioAvg"] = waitRatioAvg;
|
| }
|
| if (waitRatioMax != null) {
|
| _json["waitRatioMax"] = waitRatioMax;
|
| }
|
| + if (writeMsAvg != null) {
|
| + _json["writeMsAvg"] = writeMsAvg;
|
| + }
|
| + if (writeMsMax != null) {
|
| + _json["writeMsMax"] = writeMsMax;
|
| + }
|
| if (writeRatioAvg != null) {
|
| _json["writeRatioAvg"] = writeRatioAvg;
|
| }
|
| @@ -2226,8 +2306,8 @@ class ExternalDataConfiguration {
|
| * load jobs apply to external data sources. For Google Cloud Bigtable URIs:
|
| * Exactly one URI can be specified and it has be a fully specified and valid
|
| * HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore
|
| - * backups, exactly one URI can be specified, and it must end with
|
| - * '.backup_info'. Also, the '*' wildcard character is not allowed.
|
| + * backups, exactly one URI can be specified. Also, the '*' wildcard character
|
| + * is not allowed.
|
| */
|
| core.List<core.String> sourceUris;
|
|
|
| @@ -2864,8 +2944,9 @@ class JobConfigurationLoad {
|
| core.String schemaInlineFormat;
|
| /**
|
| * [Experimental] Allows the schema of the desitination table to be updated as
|
| - * a side effect of the load job. Schema update options are supported in two
|
| - * cases: when writeDisposition is WRITE_APPEND; when writeDisposition is
|
| + * a side effect of the load job if a schema is autodetected or supplied in
|
| + * the job configuration. Schema update options are supported in two cases:
|
| + * when writeDisposition is WRITE_APPEND; when writeDisposition is
|
| * WRITE_TRUNCATE and the destination table is a partition of a table,
|
| * specified by partition decorators. For normal tables, WRITE_TRUNCATE will
|
| * always overwrite the schema. One or more of the following values are
|
| @@ -2894,10 +2975,15 @@ class JobConfigurationLoad {
|
| * load jobs apply to external data sources. For Google Cloud Bigtable URIs:
|
| * Exactly one URI can be specified and it has be a fully specified and valid
|
| * HTTPS URL for a Google Cloud Bigtable table. For Google Cloud Datastore
|
| - * backups: Exactly one URI can be specified, and it must end with
|
| - * '.backup_info'. Also, the '*' wildcard character is not allowed.
|
| + * backups: Exactly one URI can be specified. Also, the '*' wildcard character
|
| + * is not allowed.
|
| */
|
| core.List<core.String> sourceUris;
|
| + /**
|
| + * [Experimental] If specified, configures time-based partitioning for the
|
| + * destination table.
|
| + */
|
| + TimePartitioning timePartitioning;
|
| /**
|
| * [Optional] Specifies the action that occurs if the destination table
|
| * already exists. The following values are supported: WRITE_TRUNCATE: If the
|
| @@ -2971,6 +3057,9 @@ class JobConfigurationLoad {
|
| if (_json.containsKey("sourceUris")) {
|
| sourceUris = _json["sourceUris"];
|
| }
|
| + if (_json.containsKey("timePartitioning")) {
|
| + timePartitioning = new TimePartitioning.fromJson(_json["timePartitioning"]);
|
| + }
|
| if (_json.containsKey("writeDisposition")) {
|
| writeDisposition = _json["writeDisposition"];
|
| }
|
| @@ -3035,6 +3124,9 @@ class JobConfigurationLoad {
|
| if (sourceUris != null) {
|
| _json["sourceUris"] = sourceUris;
|
| }
|
| + if (timePartitioning != null) {
|
| + _json["timePartitioning"] = (timePartitioning).toJson();
|
| + }
|
| if (writeDisposition != null) {
|
| _json["writeDisposition"] = writeDisposition;
|
| }
|
| @@ -3130,6 +3222,11 @@ class JobConfigurationQuery {
|
| * as if it were a standard BigQuery table.
|
| */
|
| core.Map<core.String, ExternalDataConfiguration> tableDefinitions;
|
| + /**
|
| + * [Experimental] If specified, configures time-based partitioning for the
|
| + * destination table.
|
| + */
|
| + TimePartitioning timePartitioning;
|
| /**
|
| * Specifies whether to use BigQuery's legacy SQL dialect for this query. The
|
| * default value is true. If set to false, the query will use BigQuery's
|
| @@ -3207,6 +3304,9 @@ class JobConfigurationQuery {
|
| if (_json.containsKey("tableDefinitions")) {
|
| tableDefinitions = commons.mapMap<core.Map<core.String, core.Object>, ExternalDataConfiguration>(_json["tableDefinitions"], (core.Map<core.String, core.Object> item) => new ExternalDataConfiguration.fromJson(item));
|
| }
|
| + if (_json.containsKey("timePartitioning")) {
|
| + timePartitioning = new TimePartitioning.fromJson(_json["timePartitioning"]);
|
| + }
|
| if (_json.containsKey("useLegacySql")) {
|
| useLegacySql = _json["useLegacySql"];
|
| }
|
| @@ -3265,6 +3365,9 @@ class JobConfigurationQuery {
|
| if (tableDefinitions != null) {
|
| _json["tableDefinitions"] = commons.mapMap<ExternalDataConfiguration, core.Map<core.String, core.Object>>(tableDefinitions, (ExternalDataConfiguration item) => (item).toJson());
|
| }
|
| + if (timePartitioning != null) {
|
| + _json["timePartitioning"] = (timePartitioning).toJson();
|
| + }
|
| if (useLegacySql != null) {
|
| _json["useLegacySql"] = useLegacySql;
|
| }
|
|
|