OLD | NEW |
1 { | 1 { |
2 "auth": { | 2 "auth": { |
3 "oauth2": { | 3 "oauth2": { |
4 "scopes": { | 4 "scopes": { |
5 "https://www.googleapis.com/auth/bigquery": { | 5 "https://www.googleapis.com/auth/bigquery": { |
6 "description": "View and manage your data in Google BigQuery
" | 6 "description": "View and manage your data in Google BigQuery
" |
7 }, | 7 }, |
8 "https://www.googleapis.com/auth/bigquery.insertdata": { | 8 "https://www.googleapis.com/auth/bigquery.insertdata": { |
9 "description": "Insert data into Google BigQuery" | 9 "description": "Insert data into Google BigQuery" |
10 }, | 10 }, |
(...skipping 11 matching lines...) Expand all Loading... |
22 } | 22 } |
23 } | 23 } |
24 } | 24 } |
25 }, | 25 }, |
26 "basePath": "/bigquery/v2/", | 26 "basePath": "/bigquery/v2/", |
27 "baseUrl": "https://www.googleapis.com/bigquery/v2/", | 27 "baseUrl": "https://www.googleapis.com/bigquery/v2/", |
28 "batchPath": "batch", | 28 "batchPath": "batch", |
29 "description": "A data platform for customers to create, manage, share and q
uery data.", | 29 "description": "A data platform for customers to create, manage, share and q
uery data.", |
30 "discoveryVersion": "v1", | 30 "discoveryVersion": "v1", |
31 "documentationLink": "https://developers.google.com/bigquery/docs/overview", | 31 "documentationLink": "https://developers.google.com/bigquery/docs/overview", |
32 "etag": "\"ye6orv2F-1npMW3u9suM3a7C5Bo/oDdLsdibHBJaKlk0IjPRWSI7TBM\"", | 32 "etag": "\"ye6orv2F-1npMW3u9suM3a7C5Bo/KlquM6UqjmIyuGctKF7ppu43Avk\"", |
33 "icons": { | 33 "icons": { |
34 "x16": "https://www.google.com/images/icons/product/search-16.gif", | 34 "x16": "https://www.google.com/images/icons/product/search-16.gif", |
35 "x32": "https://www.google.com/images/icons/product/search-32.gif" | 35 "x32": "https://www.google.com/images/icons/product/search-32.gif" |
36 }, | 36 }, |
37 "id": "bigquery:v2", | 37 "id": "bigquery:v2", |
38 "kind": "discovery#restDescription", | 38 "kind": "discovery#restDescription", |
39 "name": "bigquery", | 39 "name": "bigquery", |
40 "ownerDomain": "google.com", | 40 "ownerDomain": "google.com", |
41 "ownerName": "Google", | 41 "ownerName": "Google", |
42 "parameters": { | 42 "parameters": { |
(...skipping 839 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
882 "$ref": "Table" | 882 "$ref": "Table" |
883 }, | 883 }, |
884 "scopes": [ | 884 "scopes": [ |
885 "https://www.googleapis.com/auth/bigquery", | 885 "https://www.googleapis.com/auth/bigquery", |
886 "https://www.googleapis.com/auth/cloud-platform" | 886 "https://www.googleapis.com/auth/cloud-platform" |
887 ] | 887 ] |
888 } | 888 } |
889 } | 889 } |
890 } | 890 } |
891 }, | 891 }, |
892 "revision": "20150226", | 892 "revision": "20150326", |
893 "rootUrl": "https://www.googleapis.com/", | 893 "rootUrl": "https://www.googleapis.com/", |
894 "schemas": { | 894 "schemas": { |
| 895 "CsvOptions": { |
| 896 "id": "CsvOptions", |
| 897 "properties": { |
| 898 "allowJaggedRows": { |
| 899 "description": "[Optional] Indicates if BigQuery should acce
pt rows that are missing trailing optional columns. If true, BigQuery treats mis
sing trailing columns as null values. If false, records with missing trailing co
lumns are treated as bad records, and if there are too many bad records, an inva
lid error is returned in the job result. The default value is false.", |
| 900 "type": "boolean" |
| 901 }, |
| 902 "allowQuotedNewlines": { |
| 903 "description": "[Optional] Indicates if BigQuery should allo
w quoted data sections that contain newline characters in a CSV file. The defaul
t value is false.", |
| 904 "type": "boolean" |
| 905 }, |
| 906 "encoding": { |
| 907 "description": "[Optional] The character encoding of the dat
a. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. Big
Query decodes the data after the raw, binary data has been split using the value
s of the quote and fieldDelimiter properties.", |
| 908 "type": "string" |
| 909 }, |
| 910 "fieldDelimiter": { |
| 911 "description": "[Optional] The separator for fields in a CSV
file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the fi
rst byte of the encoded string to split the data in its raw, binary state. BigQu
ery also supports the escape sequence \"\\t\" to specify a tab separator. The de
fault value is a comma (',').", |
| 912 "type": "string" |
| 913 }, |
| 914 "quote": { |
| 915 "description": "[Optional] The value that is used to quote d
ata sections in a CSV file. BigQuery converts the string to ISO-8859-1 encoding,
and then uses the first byte of the encoded string to split the data in its raw
, binary state. The default value is a double-quote ('\"'). If your data does no
t contain quoted sections, set the property value to an empty string. If your da
ta contains quoted newline characters, you must also set the allowQuotedNewlines
property to true.", |
| 916 "type": "string" |
| 917 }, |
| 918 "skipLeadingRows": { |
| 919 "description": "[Optional] The number of rows at the top of
a CSV file that BigQuery will skip when reading the data. The default value is 0
. This property is useful if you have header rows in the file that should be ski
pped.", |
| 920 "format": "int32", |
| 921 "type": "integer" |
| 922 } |
| 923 }, |
| 924 "type": "object" |
| 925 }, |
895 "Dataset": { | 926 "Dataset": { |
896 "id": "Dataset", | 927 "id": "Dataset", |
897 "properties": { | 928 "properties": { |
898 "access": { | 929 "access": { |
899 "description": "[Optional] An array of objects that define d
ataset access for one or more entities. You can set this property when inserting
or updating a dataset in order to control who is allowed to access the data. If
unspecified at dataset creation time, BigQuery adds default dataset access for
the following entities: access.specialGroup: projectReaders; access.role: READER
; access.specialGroup: projectWriters; access.role: WRITER; access.specialGroup:
projectOwners; access.role: OWNER; access.userByEmail: [dataset creator email];
access.role: OWNER;", | 930 "description": "[Optional] An array of objects that define d
ataset access for one or more entities. You can set this property when inserting
or updating a dataset in order to control who is allowed to access the data. If
unspecified at dataset creation time, BigQuery adds default dataset access for
the following entities: access.specialGroup: projectReaders; access.role: READER
; access.specialGroup: projectWriters; access.role: WRITER; access.specialGroup:
projectOwners; access.role: OWNER; access.userByEmail: [dataset creator email];
access.role: OWNER;", |
900 "items": { | 931 "items": { |
901 "properties": { | 932 "properties": { |
902 "domain": { | 933 "domain": { |
903 "description": "[Pick one] A domain to grant acc
ess to. Any users signed in with the domain specified will be granted the specif
ied access. Example: \"example.com\".", | 934 "description": "[Pick one] A domain to grant acc
ess to. Any users signed in with the domain specified will be granted the specif
ied access. Example: \"example.com\".", |
904 "type": "string" | 935 "type": "string" |
(...skipping 25 matching lines...) Expand all Loading... |
930 }, | 961 }, |
931 "creationTime": { | 962 "creationTime": { |
932 "description": "[Output-only] The time when this dataset was
created, in milliseconds since the epoch.", | 963 "description": "[Output-only] The time when this dataset was
created, in milliseconds since the epoch.", |
933 "format": "int64", | 964 "format": "int64", |
934 "type": "string" | 965 "type": "string" |
935 }, | 966 }, |
936 "datasetReference": { | 967 "datasetReference": { |
937 "$ref": "DatasetReference", | 968 "$ref": "DatasetReference", |
938 "description": "[Required] A reference that identifies the d
ataset." | 969 "description": "[Required] A reference that identifies the d
ataset." |
939 }, | 970 }, |
| 971 "defaultTableExpirationMs": { |
| 972 "description": "[Experimental] The default lifetime of all t
ables in the dataset, in milliseconds. The minimum value is 3600000 milliseconds
(one hour). Once this property is set, all newly-created tables in the dataset
will have an expirationTime property set to the creation time plus the value in
this property, and changing the value will only affect new tables, not existing
ones. When the expirationTime for a given table is reached, that table will be d
eleted automatically. If a table's expirationTime is modified or removed before
the table expires, or if you provide an explicit expirationTime when creating a
table, that value takes precedence over the default expiration time indicated by
this property.", |
| 973 "format": "int64", |
| 974 "type": "string" |
| 975 }, |
940 "description": { | 976 "description": { |
941 "description": "[Optional] A user-friendly description of th
e dataset.", | 977 "description": "[Optional] A user-friendly description of th
e dataset.", |
942 "type": "string" | 978 "type": "string" |
943 }, | 979 }, |
944 "etag": { | 980 "etag": { |
945 "description": "[Output-only] A hash of the resource.", | 981 "description": "[Output-only] A hash of the resource.", |
946 "type": "string" | 982 "type": "string" |
947 }, | 983 }, |
948 "friendlyName": { | 984 "friendlyName": { |
949 "description": "[Optional] A descriptive name for the datase
t.", | 985 "description": "[Optional] A descriptive name for the datase
t.", |
(...skipping 104 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1054 "description": "A human-readable description of the error.", | 1090 "description": "A human-readable description of the error.", |
1055 "type": "string" | 1091 "type": "string" |
1056 }, | 1092 }, |
1057 "reason": { | 1093 "reason": { |
1058 "description": "A short error code that summarizes the error
.", | 1094 "description": "A short error code that summarizes the error
.", |
1059 "type": "string" | 1095 "type": "string" |
1060 } | 1096 } |
1061 }, | 1097 }, |
1062 "type": "object" | 1098 "type": "object" |
1063 }, | 1099 }, |
| 1100 "ExternalDataConfiguration": { |
| 1101 "id": "ExternalDataConfiguration", |
| 1102 "properties": { |
| 1103 "compression": { |
| 1104 "description": "[Optional] The compression type of the data
source. Possible values include GZIP and NONE. The default value is NONE.", |
| 1105 "type": "string" |
| 1106 }, |
| 1107 "csvOptions": { |
| 1108 "$ref": "CsvOptions", |
| 1109 "description": "Additional properties to set if sourceFormat
is set to CSV." |
| 1110 }, |
| 1111 "ignoreUnknownValues": { |
| 1112 "description": "[Optional] Indicates if BigQuery should allo
w extra values that are not represented in the table schema. If true, the extra
values are ignored. If false, records with extra columns are treated as bad reco
rds, and if there are too many bad records, an invalid error is returned in the
job result. The default value is false. The sourceFormat property determines wha
t BigQuery treats as an extra value: CSV: Trailing columns", |
| 1113 "type": "boolean" |
| 1114 }, |
| 1115 "maxBadRecords": { |
| 1116 "description": "[Optional] The maximum number of bad records
that BigQuery can ignore when reading data. If the number of bad records exceed
s this value, an invalid error is returned in the job result. The default value
is 0, which requires that all records are valid.", |
| 1117 "format": "int32", |
| 1118 "type": "integer" |
| 1119 }, |
| 1120 "schema": { |
| 1121 "$ref": "TableSchema", |
| 1122 "description": "[Required] The schema for the data." |
| 1123 }, |
| 1124 "sourceFormat": { |
| 1125 "description": "[Optional] The data format. External data so
urces must be in CSV format. The default value is CSV.", |
| 1126 "type": "string" |
| 1127 }, |
| 1128 "sourceUris": { |
| 1129 "description": "[Required] The fully-qualified URIs that poi
nt to your data in Google Cloud Storage. Each URI can contain one '*' wildcard c
haracter and it must come after the 'bucket' name. CSV limits related to load jo
bs apply to external data sources, plus an additional limit of 10 GB maximum siz
e across all URIs.", |
| 1130 "items": { |
| 1131 "type": "string" |
| 1132 }, |
| 1133 "type": "array" |
| 1134 } |
| 1135 }, |
| 1136 "type": "object" |
| 1137 }, |
1064 "GetQueryResultsResponse": { | 1138 "GetQueryResultsResponse": { |
1065 "id": "GetQueryResultsResponse", | 1139 "id": "GetQueryResultsResponse", |
1066 "properties": { | 1140 "properties": { |
1067 "cacheHit": { | 1141 "cacheHit": { |
1068 "description": "Whether the query result was fetched from th
e query cache.", | 1142 "description": "Whether the query result was fetched from th
e query cache.", |
1069 "type": "boolean" | 1143 "type": "boolean" |
1070 }, | 1144 }, |
1071 "etag": { | 1145 "etag": { |
1072 "description": "A hash of this response.", | 1146 "description": "A hash of this response.", |
1073 "type": "string" | 1147 "type": "string" |
(...skipping 171 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1245 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", | 1319 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", |
1246 "type": "string" | 1320 "type": "string" |
1247 } | 1321 } |
1248 }, | 1322 }, |
1249 "type": "object" | 1323 "type": "object" |
1250 }, | 1324 }, |
1251 "JobConfigurationLoad": { | 1325 "JobConfigurationLoad": { |
1252 "id": "JobConfigurationLoad", | 1326 "id": "JobConfigurationLoad", |
1253 "properties": { | 1327 "properties": { |
1254 "allowJaggedRows": { | 1328 "allowJaggedRows": { |
1255 "description": "[Optional] Accept rows that are missing trai
ling optional columns. The missing values are treated as nulls. Default is false
which treats short rows as errors. Only applicable to CSV, ignored for other fo
rmats.", | 1329 "description": "[Optional] Accept rows that are missing trai
ling optional columns. The missing values are treated as nulls. If false, record
s with missing trailing columns are treated as bad records, and if there are too
many bad records, an invalid error is returned in the job result. The default v
alue is false. Only applicable to CSV, ignored for other formats.", |
1256 "type": "boolean" | 1330 "type": "boolean" |
1257 }, | 1331 }, |
1258 "allowQuotedNewlines": { | 1332 "allowQuotedNewlines": { |
1259 "description": "Indicates if BigQuery should allow quoted da
ta sections that contain newline characters in a CSV file. The default value is
false.", | 1333 "description": "Indicates if BigQuery should allow quoted da
ta sections that contain newline characters in a CSV file. The default value is
false.", |
1260 "type": "boolean" | 1334 "type": "boolean" |
1261 }, | 1335 }, |
1262 "createDisposition": { | 1336 "createDisposition": { |
1263 "description": "[Optional] Specifies whether the job is allo
wed to create new tables. The following values are supported: CREATE_IF_NEEDED:
If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table
must already exist. If it does not, a 'notFound' error is returned in the job r
esult. The default value is CREATE_IF_NEEDED. Creation, truncation and append ac
tions occur as one atomic update upon job completion.", | 1337 "description": "[Optional] Specifies whether the job is allo
wed to create new tables. The following values are supported: CREATE_IF_NEEDED:
If the table does not exist, BigQuery creates the table. CREATE_NEVER: The table
must already exist. If it does not, a 'notFound' error is returned in the job r
esult. The default value is CREATE_IF_NEEDED. Creation, truncation and append ac
tions occur as one atomic update upon job completion.", |
1264 "type": "string" | 1338 "type": "string" |
1265 }, | 1339 }, |
1266 "destinationTable": { | 1340 "destinationTable": { |
1267 "$ref": "TableReference", | 1341 "$ref": "TableReference", |
1268 "description": "[Required] The destination table to load the
data into." | 1342 "description": "[Required] The destination table to load the
data into." |
1269 }, | 1343 }, |
1270 "encoding": { | 1344 "encoding": { |
1271 "description": "[Optional] The character encoding of the dat
a. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. Big
Query decodes the data after the raw, binary data has been split using the value
s of the quote and fieldDelimiter properties.", | 1345 "description": "[Optional] The character encoding of the dat
a. The supported values are UTF-8 or ISO-8859-1. The default value is UTF-8. Big
Query decodes the data after the raw, binary data has been split using the value
s of the quote and fieldDelimiter properties.", |
1272 "type": "string" | 1346 "type": "string" |
1273 }, | 1347 }, |
1274 "fieldDelimiter": { | 1348 "fieldDelimiter": { |
1275 "description": "[Optional] The separator for fields in a CSV
file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the fi
rst byte of the encoded string to split the data in its raw, binary state. BigQu
ery also supports the escape sequence \"\\t\" to specify a tab separator. The de
fault value is a comma (',').", | 1349 "description": "[Optional] The separator for fields in a CSV
file. BigQuery converts the string to ISO-8859-1 encoding, and then uses the fi
rst byte of the encoded string to split the data in its raw, binary state. BigQu
ery also supports the escape sequence \"\\t\" to specify a tab separator. The de
fault value is a comma (',').", |
1276 "type": "string" | 1350 "type": "string" |
1277 }, | 1351 }, |
1278 "ignoreUnknownValues": { | 1352 "ignoreUnknownValues": { |
1279 "description": "[Optional] Accept rows that contain values t
hat do not match the schema. The unknown values are ignored. Default is false wh
ich treats unknown values as errors. For CSV this ignores extra values at the en
d of a line. For JSON this ignores named values that do not match any column nam
e.", | 1353 "description": "[Optional] Indicates if BigQuery should allo
w extra values that are not represented in the table schema. If true, the extra
values are ignored. If false, records with extra columns are treated as bad reco
rds, and if there are too many bad records, an invalid error is returned in the
job result. The default value is false. The sourceFormat property determines wha
t BigQuery treats as an extra value: CSV: Trailing columns JSON: Named values th
at don't match any column names", |
1280 "type": "boolean" | 1354 "type": "boolean" |
1281 }, | 1355 }, |
1282 "maxBadRecords": { | 1356 "maxBadRecords": { |
1283 "description": "[Optional] The maximum number of bad records
that BigQuery can ignore when running the job. If the number of bad records exc
eeds this value, an 'invalid' error is returned in the job result and the job fa
ils. The default value is 0, which requires that all records are valid.", | 1357 "description": "[Optional] The maximum number of bad records
that BigQuery can ignore when running the job. If the number of bad records exc
eeds this value, an invalid error is returned in the job result. The default val
ue is 0, which requires that all records are valid.", |
1284 "format": "int32", | 1358 "format": "int32", |
1285 "type": "integer" | 1359 "type": "integer" |
1286 }, | 1360 }, |
1287 "projectionFields": { | 1361 "projectionFields": { |
1288 "description": "[Experimental] Names(case-sensitive) of prop
erties to keep when importing data. If this is populated, only the specified pro
perties will be imported for each entity. Currently, this is only supported for
DATASTORE_BACKUP imports and only top level properties are supported. If any spe
cified property is not found in the Datastore 'Kind' being imported, that is an
error. Note: This feature is experimental and can change in the future.", | 1362 "description": "[Experimental] Names(case-sensitive) of prop
erties to keep when importing data. If this is populated, only the specified pro
perties will be imported for each entity. Currently, this is only supported for
DATASTORE_BACKUP imports and only top level properties are supported. If any spe
cified property is not found in the Datastore 'Kind' being imported, that is an
error. Note: This feature is experimental and can change in the future.", |
1289 "items": { | 1363 "items": { |
1290 "type": "string" | 1364 "type": "string" |
1291 }, | 1365 }, |
1292 "type": "array" | 1366 "type": "array" |
1293 }, | 1367 }, |
(...skipping 16 matching lines...) Expand all Loading... |
1310 "skipLeadingRows": { | 1384 "skipLeadingRows": { |
1311 "description": "[Optional] The number of rows at the top of
a CSV file that BigQuery will skip when loading the data. The default value is 0
. This property is useful if you have header rows in the file that should be ski
pped.", | 1385 "description": "[Optional] The number of rows at the top of
a CSV file that BigQuery will skip when loading the data. The default value is 0
. This property is useful if you have header rows in the file that should be ski
pped.", |
1312 "format": "int32", | 1386 "format": "int32", |
1313 "type": "integer" | 1387 "type": "integer" |
1314 }, | 1388 }, |
1315 "sourceFormat": { | 1389 "sourceFormat": { |
1316 "description": "[Optional] The format of the data files. For
CSV files, specify \"CSV\". For datastore backups, specify \"DATASTORE_BACKUP\"
. For newline-delimited JSON, specify \"NEWLINE_DELIMITED_JSON\". The default va
lue is CSV.", | 1390 "description": "[Optional] The format of the data files. For
CSV files, specify \"CSV\". For datastore backups, specify \"DATASTORE_BACKUP\"
. For newline-delimited JSON, specify \"NEWLINE_DELIMITED_JSON\". The default va
lue is CSV.", |
1317 "type": "string" | 1391 "type": "string" |
1318 }, | 1392 }, |
1319 "sourceUris": { | 1393 "sourceUris": { |
1320 "description": "[Required] The fully-qualified URIs that poi
nt to your data in Google Cloud Storage. Wildcard names are only supported when
they appear at the end of the URI.", | 1394 "description": "[Required] The fully-qualified URIs that poi
nt to your data in Google Cloud Storage. Each URI can contain one '*' wildcard c
haracter and it must come after the 'bucket' name.", |
1321 "items": { | 1395 "items": { |
1322 "type": "string" | 1396 "type": "string" |
1323 }, | 1397 }, |
1324 "type": "array" | 1398 "type": "array" |
1325 }, | 1399 }, |
1326 "writeDisposition": { | 1400 "writeDisposition": { |
1327 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", | 1401 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", |
1328 "type": "string" | 1402 "type": "string" |
1329 } | 1403 } |
1330 }, | 1404 }, |
(...skipping 27 matching lines...) Expand all Loading... |
1358 "type": "boolean" | 1432 "type": "boolean" |
1359 }, | 1433 }, |
1360 "priority": { | 1434 "priority": { |
1361 "description": "[Optional] Specifies a priority for the quer
y. Possible values include INTERACTIVE and BATCH. The default value is INTERACTI
VE.", | 1435 "description": "[Optional] Specifies a priority for the quer
y. Possible values include INTERACTIVE and BATCH. The default value is INTERACTI
VE.", |
1362 "type": "string" | 1436 "type": "string" |
1363 }, | 1437 }, |
1364 "query": { | 1438 "query": { |
1365 "description": "[Required] BigQuery SQL query to execute.", | 1439 "description": "[Required] BigQuery SQL query to execute.", |
1366 "type": "string" | 1440 "type": "string" |
1367 }, | 1441 }, |
| 1442 "tableDefinitions": { |
| 1443 "additionalProperties": { |
| 1444 "$ref": "ExternalDataConfiguration" |
| 1445 }, |
| 1446 "description": "[Experimental] If querying an external data
source outside of BigQuery, describes the data format, location and other proper
ties of the data source. By defining these properties, the data source can then
be queried as if it were a standard BigQuery table.", |
| 1447 "type": "object" |
| 1448 }, |
1368 "useQueryCache": { | 1449 "useQueryCache": { |
1369 "description": "[Optional] Whether to look for the result in
the query cache. The query cache is a best-effort cache that will be flushed wh
enever tables in the query are modified. Moreover, the query cache is only avail
able when a query does not have a destination table specified.", | 1450 "description": "[Optional] Whether to look for the result in
the query cache. The query cache is a best-effort cache that will be flushed wh
enever tables in the query are modified. Moreover, the query cache is only avail
able when a query does not have a destination table specified.", |
1370 "type": "boolean" | 1451 "type": "boolean" |
1371 }, | 1452 }, |
1372 "writeDisposition": { | 1453 "writeDisposition": { |
1373 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", | 1454 "description": "[Optional] Specifies the action that occurs
if the destination table already exists. The following values are supported: WRI
TE_TRUNCATE: If the table already exists, BigQuery overwrites the table data. WR
ITE_APPEND: If the table already exists, BigQuery appends the data to the table.
WRITE_EMPTY: If the table already exists and contains data, a 'duplicate' error
is returned in the job result. The default value is WRITE_EMPTY. Each action is
atomic and only occurs if BigQuery is able to complete the job successfully. Cr
eation, truncation and append actions occur as one atomic update upon job comple
tion.", |
1374 "type": "string" | 1455 "type": "string" |
1375 } | 1456 } |
1376 }, | 1457 }, |
1377 "type": "object" | 1458 "type": "object" |
(...skipping 484 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
1862 "properties": { | 1943 "properties": { |
1863 "v": { | 1944 "v": { |
1864 "type": "any" | 1945 "type": "any" |
1865 } | 1946 } |
1866 }, | 1947 }, |
1867 "type": "object" | 1948 "type": "object" |
1868 }, | 1949 }, |
1869 "TableDataInsertAllRequest": { | 1950 "TableDataInsertAllRequest": { |
1870 "id": "TableDataInsertAllRequest", | 1951 "id": "TableDataInsertAllRequest", |
1871 "properties": { | 1952 "properties": { |
| 1953 "ignoreUnknownValues": { |
| 1954 "description": "[Optional] Accept rows that contain values t
hat do not match the schema. The unknown values are ignored. Default is false, w
hich treats unknown values as errors.", |
| 1955 "type": "boolean" |
| 1956 }, |
1872 "kind": { | 1957 "kind": { |
1873 "default": "bigquery#tableDataInsertAllRequest", | 1958 "default": "bigquery#tableDataInsertAllRequest", |
1874 "description": "The resource type of the response.", | 1959 "description": "The resource type of the response.", |
1875 "type": "string" | 1960 "type": "string" |
1876 }, | 1961 }, |
1877 "rows": { | 1962 "rows": { |
1878 "description": "The rows to insert.", | 1963 "description": "The rows to insert.", |
1879 "items": { | 1964 "items": { |
1880 "properties": { | 1965 "properties": { |
1881 "insertId": { | 1966 "insertId": { |
1882 "description": "[Optional] A unique ID for each
row. BigQuery uses this property to detect duplicate insertion requests on a bes
t-effort basis.", | 1967 "description": "[Optional] A unique ID for each
row. BigQuery uses this property to detect duplicate insertion requests on a bes
t-effort basis.", |
1883 "type": "string" | 1968 "type": "string" |
1884 }, | 1969 }, |
1885 "json": { | 1970 "json": { |
1886 "$ref": "JsonObject", | 1971 "$ref": "JsonObject", |
1887 "description": "[Required] A JSON object that co
ntains a row of data. The object's properties and values must match the destinat
ion table's schema." | 1972 "description": "[Required] A JSON object that co
ntains a row of data. The object's properties and values must match the destinat
ion table's schema." |
1888 } | 1973 } |
1889 }, | 1974 }, |
1890 "type": "object" | 1975 "type": "object" |
1891 }, | 1976 }, |
1892 "type": "array" | 1977 "type": "array" |
| 1978 }, |
| 1979 "skipInvalidRows": { |
| 1980 "description": "[Optional] Insert all valid rows of a reques
t, even if invalid rows exist. The default value is false, which causes the enti
re request to fail if any invalid rows exist.", |
| 1981 "type": "boolean" |
1893 } | 1982 } |
1894 }, | 1983 }, |
1895 "type": "object" | 1984 "type": "object" |
1896 }, | 1985 }, |
1897 "TableDataInsertAllResponse": { | 1986 "TableDataInsertAllResponse": { |
1898 "id": "TableDataInsertAllResponse", | 1987 "id": "TableDataInsertAllResponse", |
1899 "properties": { | 1988 "properties": { |
1900 "insertErrors": { | 1989 "insertErrors": { |
1901 "description": "An array of errors for rows that were not in
serted.", | 1990 "description": "An array of errors for rows that were not in
serted.", |
1902 "items": { | 1991 "items": { |
(...skipping 204 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
2107 "type": "string" | 2196 "type": "string" |
2108 } | 2197 } |
2109 }, | 2198 }, |
2110 "type": "object" | 2199 "type": "object" |
2111 } | 2200 } |
2112 }, | 2201 }, |
2113 "servicePath": "bigquery/v2/", | 2202 "servicePath": "bigquery/v2/", |
2114 "title": "BigQuery API", | 2203 "title": "BigQuery API", |
2115 "version": "v2" | 2204 "version": "v2" |
2116 } | 2205 } |
OLD | NEW |