Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(107)

Side by Side Diff: discovery/googleapis_beta/speech__v1beta1.json

Issue 2779563003: Api-roll 47: 2017-03-27 (Closed)
Patch Set: Created 3 years, 8 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View unified diff | Download patch
OLDNEW
1 { 1 {
2 "auth": { 2 "auth": {
3 "oauth2": { 3 "oauth2": {
4 "scopes": { 4 "scopes": {
5 "https://www.googleapis.com/auth/cloud-platform": { 5 "https://www.googleapis.com/auth/cloud-platform": {
6 "description": "View and manage your data across Google Clou d Platform services" 6 "description": "View and manage your data across Google Clou d Platform services"
7 } 7 }
8 } 8 }
9 } 9 }
10 }, 10 },
11 "basePath": "", 11 "basePath": "",
12 "baseUrl": "https://speech.googleapis.com/", 12 "baseUrl": "https://speech.googleapis.com/",
13 "batchPath": "batch", 13 "batchPath": "batch",
14 "description": "Google Cloud Speech API.", 14 "description": "Google Cloud Speech API.",
15 "discoveryVersion": "v1", 15 "discoveryVersion": "v1",
16 "documentationLink": "https://cloud.google.com/speech/", 16 "documentationLink": "https://cloud.google.com/speech/",
17 "icons": { 17 "icons": {
18 "x16": "http://www.google.com/images/icons/product/search-16.gif", 18 "x16": "http://www.google.com/images/icons/product/search-16.gif",
19 "x32": "http://www.google.com/images/icons/product/search-32.gif" 19 "x32": "http://www.google.com/images/icons/product/search-32.gif"
20 }, 20 },
21 "id": "speech:v1beta1", 21 "id": "speech:v1beta1",
22 "kind": "discovery#restDescription", 22 "kind": "discovery#restDescription",
23 "name": "speech", 23 "name": "speech",
24 "ownerDomain": "google.com", 24 "ownerDomain": "google.com",
25 "ownerName": "Google", 25 "ownerName": "Google",
26 "parameters": { 26 "parameters": {
27 "oauth_token": {
28 "description": "OAuth 2.0 token for the current user.",
29 "location": "query",
30 "type": "string"
31 },
32 "bearer_token": {
33 "description": "OAuth bearer token.",
34 "location": "query",
35 "type": "string"
36 },
27 "upload_protocol": { 37 "upload_protocol": {
28 "description": "Upload protocol for media (e.g. \"raw\", \"multipart \").", 38 "description": "Upload protocol for media (e.g. \"raw\", \"multipart \").",
29 "location": "query", 39 "location": "query",
30 "type": "string" 40 "type": "string"
31 }, 41 },
32 "prettyPrint": { 42 "prettyPrint": {
33 "default": "true", 43 "default": "true",
34 "description": "Returns response with indentations and line breaks." , 44 "description": "Returns response with indentations and line breaks." ,
35 "location": "query", 45 "location": "query",
36 "type": "boolean" 46 "type": "boolean"
37 }, 47 },
48 "fields": {
49 "description": "Selector specifying which fields to include in a par tial response.",
50 "location": "query",
51 "type": "string"
52 },
38 "uploadType": { 53 "uploadType": {
39 "description": "Legacy upload protocol for media (e.g. \"media\", \" multipart\").", 54 "description": "Legacy upload protocol for media (e.g. \"media\", \" multipart\").",
40 "location": "query", 55 "location": "query",
41 "type": "string" 56 "type": "string"
42 }, 57 },
43 "fields": {
44 "description": "Selector specifying which fields to include in a par tial response.",
45 "location": "query",
46 "type": "string"
47 },
48 "callback": {
49 "description": "JSONP",
50 "location": "query",
51 "type": "string"
52 },
53 "$.xgafv": { 58 "$.xgafv": {
54 "description": "V1 error format.", 59 "description": "V1 error format.",
55 "enum": [ 60 "enum": [
56 "1", 61 "1",
57 "2" 62 "2"
58 ], 63 ],
59 "enumDescriptions": [ 64 "enumDescriptions": [
60 "v1 error format", 65 "v1 error format",
61 "v2 error format" 66 "v2 error format"
62 ], 67 ],
63 "location": "query", 68 "location": "query",
64 "type": "string" 69 "type": "string"
65 }, 70 },
71 "callback": {
72 "description": "JSONP",
73 "location": "query",
74 "type": "string"
75 },
66 "alt": { 76 "alt": {
67 "default": "json", 77 "default": "json",
68 "description": "Data format for response.", 78 "description": "Data format for response.",
69 "enum": [ 79 "enum": [
70 "json", 80 "json",
71 "media", 81 "media",
72 "proto" 82 "proto"
73 ], 83 ],
74 "enumDescriptions": [ 84 "enumDescriptions": [
75 "Responses with Content-Type of application/json", 85 "Responses with Content-Type of application/json",
76 "Media download with context-dependent Content-Type", 86 "Media download with context-dependent Content-Type",
77 "Responses with Content-Type of application/x-protobuf" 87 "Responses with Content-Type of application/x-protobuf"
78 ], 88 ],
79 "location": "query", 89 "location": "query",
80 "type": "string" 90 "type": "string"
81 }, 91 },
92 "access_token": {
93 "description": "OAuth access token.",
94 "location": "query",
95 "type": "string"
96 },
82 "key": { 97 "key": {
83 "description": "API key. Your API key identifies your project and pr ovides you with API access, quota, and reports. Required unless you provide an O Auth 2.0 token.", 98 "description": "API key. Your API key identifies your project and pr ovides you with API access, quota, and reports. Required unless you provide an O Auth 2.0 token.",
84 "location": "query", 99 "location": "query",
85 "type": "string" 100 "type": "string"
86 }, 101 },
87 "access_token": {
88 "description": "OAuth access token.",
89 "location": "query",
90 "type": "string"
91 },
92 "quotaUser": { 102 "quotaUser": {
93 "description": "Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exc eed 40 characters.", 103 "description": "Available to use for quota purposes for server-side applications. Can be any arbitrary string assigned to a user, but should not exc eed 40 characters.",
94 "location": "query", 104 "location": "query",
95 "type": "string" 105 "type": "string"
96 }, 106 },
97 "pp": { 107 "pp": {
98 "default": "true", 108 "default": "true",
99 "description": "Pretty-print response.", 109 "description": "Pretty-print response.",
100 "location": "query", 110 "location": "query",
101 "type": "boolean" 111 "type": "boolean"
102 },
103 "bearer_token": {
104 "description": "OAuth bearer token.",
105 "location": "query",
106 "type": "string"
107 },
108 "oauth_token": {
109 "description": "OAuth 2.0 token for the current user.",
110 "location": "query",
111 "type": "string"
112 } 112 }
113 }, 113 },
114 "protocol": "rest", 114 "protocol": "rest",
115 "resources": { 115 "resources": {
116 "operations": { 116 "operations": {
117 "methods": { 117 "methods": {
118 "cancel": { 118 "cancel": {
119 "description": "Starts asynchronous cancellation on a long-r unning operation. The server\nmakes a best effort to cancel the operation, but success is not\nguaranteed. If the server doesn't support this method, it retur ns\n`google.rpc.Code.UNIMPLEMENTED`. Clients can use\nOperations.GetOperation o r\nother methods to check whether the cancellation succeeded or whether the\nope ration completed despite cancellation. On successful cancellation,\nthe operatio n is not deleted; instead, it becomes an operation with\nan Operation.error valu e with a google.rpc.Status.code of 1,\ncorresponding to `Code.CANCELLED`.", 119 "description": "Starts asynchronous cancellation on a long-r unning operation. The server\nmakes a best effort to cancel the operation, but success is not\nguaranteed. If the server doesn't support this method, it retur ns\n`google.rpc.Code.UNIMPLEMENTED`. Clients can use\nOperations.GetOperation o r\nother methods to check whether the cancellation succeeded or whether the\nope ration completed despite cancellation. On successful cancellation,\nthe operatio n is not deleted; instead, it becomes an operation with\nan Operation.error valu e with a google.rpc.Status.code of 1,\ncorresponding to `Code.CANCELLED`.",
120 "httpMethod": "POST", 120 "httpMethod": "POST",
121 "id": "speech.operations.cancel", 121 "id": "speech.operations.cancel",
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after
162 "scopes": [ 162 "scopes": [
163 "https://www.googleapis.com/auth/cloud-platform" 163 "https://www.googleapis.com/auth/cloud-platform"
164 ] 164 ]
165 }, 165 },
166 "list": { 166 "list": {
167 "description": "Lists operations that match the specified fi lter in the request. If the\nserver doesn't support this method, it returns `UNI MPLEMENTED`.\n\nNOTE: the `name` binding below allows API services to override t he binding\nto use different resource name schemes, such as `users/*/operations` .", 167 "description": "Lists operations that match the specified fi lter in the request. If the\nserver doesn't support this method, it returns `UNI MPLEMENTED`.\n\nNOTE: the `name` binding below allows API services to override t he binding\nto use different resource name schemes, such as `users/*/operations` .",
168 "httpMethod": "GET", 168 "httpMethod": "GET",
169 "id": "speech.operations.list", 169 "id": "speech.operations.list",
170 "parameterOrder": [], 170 "parameterOrder": [],
171 "parameters": { 171 "parameters": {
172 "filter": {
173 "description": "The standard list filter.",
174 "location": "query",
175 "type": "string"
176 },
177 "name": { 172 "name": {
178 "description": "The name of the operation collection .", 173 "description": "The name of the operation collection .",
179 "location": "query", 174 "location": "query",
180 "type": "string" 175 "type": "string"
181 }, 176 },
182 "pageToken": { 177 "pageToken": {
183 "description": "The standard list page token.", 178 "description": "The standard list page token.",
184 "location": "query", 179 "location": "query",
185 "type": "string" 180 "type": "string"
186 }, 181 },
187 "pageSize": { 182 "pageSize": {
188 "description": "The standard list page size.", 183 "description": "The standard list page size.",
189 "format": "int32", 184 "format": "int32",
190 "location": "query", 185 "location": "query",
191 "type": "integer" 186 "type": "integer"
187 },
188 "filter": {
189 "description": "The standard list filter.",
190 "location": "query",
191 "type": "string"
192 } 192 }
193 }, 193 },
194 "path": "v1beta1/operations", 194 "path": "v1beta1/operations",
195 "response": { 195 "response": {
196 "$ref": "ListOperationsResponse" 196 "$ref": "ListOperationsResponse"
197 }, 197 },
198 "scopes": [ 198 "scopes": [
199 "https://www.googleapis.com/auth/cloud-platform" 199 "https://www.googleapis.com/auth/cloud-platform"
200 ] 200 ]
201 }, 201 },
(...skipping 55 matching lines...) Expand 10 before | Expand all | Expand 10 after
257 "response": { 257 "response": {
258 "$ref": "SyncRecognizeResponse" 258 "$ref": "SyncRecognizeResponse"
259 }, 259 },
260 "scopes": [ 260 "scopes": [
261 "https://www.googleapis.com/auth/cloud-platform" 261 "https://www.googleapis.com/auth/cloud-platform"
262 ] 262 ]
263 } 263 }
264 } 264 }
265 } 265 }
266 }, 266 },
267 "revision": "20170303", 267 "revision": "20170323",
268 "rootUrl": "https://speech.googleapis.com/", 268 "rootUrl": "https://speech.googleapis.com/",
269 "schemas": { 269 "schemas": {
270 "RecognitionAudio": {
271 "description": "Contains audio data in the encoding specified in the `RecognitionConfig`.\nEither `content` or `uri` must be supplied. Supplying bot h or neither\nreturns google.rpc.Code.INVALID_ARGUMENT. See\n[audio limits](http s://cloud.google.com/speech/limits#content).",
272 "id": "RecognitionAudio",
273 "properties": {
274 "content": {
275 "description": "The audio data bytes encoded as specified in \n`RecognitionConfig`. Note: as with all bytes fields, protobuffers use a\npure binary representation, whereas JSON representations use base64.",
276 "format": "byte",
277 "type": "string"
278 },
279 "uri": {
280 "description": "URI that points to a file that contains audi o data bytes as specified in\n`RecognitionConfig`. Currently, only Google Cloud Storage URIs are\nsupported, which must be specified in the following format:\n` gs://bucket_name/object_name` (other URI formats return\ngoogle.rpc.Code.INVALID _ARGUMENT). For more information, see\n[Request URIs](https://cloud.google.com/s torage/docs/reference-uris).",
281 "type": "string"
282 }
283 },
284 "type": "object"
285 },
286 "AsyncRecognizeRequest": {
287 "description": "The top-level message sent by the client for the `As yncRecognize` method.",
288 "id": "AsyncRecognizeRequest",
289 "properties": {
290 "config": {
291 "$ref": "RecognitionConfig",
292 "description": "*Required* Provides information to the recog nizer that specifies how to\nprocess the request."
293 },
294 "audio": {
295 "$ref": "RecognitionAudio",
296 "description": "*Required* The audio data to be recognized."
297 }
298 },
299 "type": "object"
300 },
270 "Operation": { 301 "Operation": {
271 "description": "This resource represents a long-running operation th at is the result of a\nnetwork API call.", 302 "description": "This resource represents a long-running operation th at is the result of a\nnetwork API call.",
272 "id": "Operation", 303 "id": "Operation",
273 "properties": { 304 "properties": {
274 "done": { 305 "done": {
275 "description": "If the value is `false`, it means the operat ion is still in progress.\nIf true, the operation is completed, and either `erro r` or `response` is\navailable.", 306 "description": "If the value is `false`, it means the operat ion is still in progress.\nIf true, the operation is completed, and either `erro r` or `response` is\navailable.",
276 "type": "boolean" 307 "type": "boolean"
277 }, 308 },
278 "response": { 309 "response": {
279 "additionalProperties": { 310 "additionalProperties": {
(...skipping 19 matching lines...) Expand all
299 "description": "Service-specific metadata associated with th e operation. It typically\ncontains progress information and common metadata su ch as create time.\nSome services might not provide such metadata. Any method t hat returns a\nlong-running operation should document the metadata type, if any. ", 330 "description": "Service-specific metadata associated with th e operation. It typically\ncontains progress information and common metadata su ch as create time.\nSome services might not provide such metadata. Any method t hat returns a\nlong-running operation should document the metadata type, if any. ",
300 "type": "object" 331 "type": "object"
301 } 332 }
302 }, 333 },
303 "type": "object" 334 "type": "object"
304 }, 335 },
305 "RecognitionConfig": { 336 "RecognitionConfig": {
306 "description": "Provides information to the recognizer that specifie s how to process the\nrequest.", 337 "description": "Provides information to the recognizer that specifie s how to process the\nrequest.",
307 "id": "RecognitionConfig", 338 "id": "RecognitionConfig",
308 "properties": { 339 "properties": {
309 "maxAlternatives": {
310 "description": "*Optional* Maximum number of recognition hyp otheses to be returned.\nSpecifically, the maximum number of `SpeechRecognitionA lternative` messages\nwithin each `SpeechRecognitionResult`.\nThe server may ret urn fewer than `max_alternatives`.\nValid values are `0`-`30`. A value of `0` or `1` will return a maximum of\none. If omitted, will return a maximum of one.",
311 "format": "int32",
312 "type": "integer"
313 },
314 "sampleRate": {
315 "description": "*Required* Sample rate in Hertz of the audio data sent in all\n`RecognitionAudio` messages. Valid values are: 8000-48000.\n1 6000 is optimal. For best results, set the sampling rate of the audio\nsource to 16000 Hz. If that's not possible, use the native sample rate of\nthe audio sour ce (instead of re-sampling).",
316 "format": "int32",
317 "type": "integer"
318 },
319 "languageCode": {
320 "description": "*Optional* The language of the supplied audi o as a BCP-47 language tag.\nExample: \"en-GB\" https://www.rfc-editor.org/rfc/ bcp/bcp47.txt\nIf omitted, defaults to \"en-US\". See\n[Language Support](https: //cloud.google.com/speech/docs/languages)\nfor a list of the currently supported language codes.",
321 "type": "string"
322 },
323 "speechContext": { 340 "speechContext": {
324 "$ref": "SpeechContext", 341 "$ref": "SpeechContext",
325 "description": "*Optional* A means to provide context to ass ist the speech recognition." 342 "description": "*Optional* A means to provide context to ass ist the speech recognition."
326 }, 343 },
327 "encoding": { 344 "encoding": {
328 "description": "*Required* Encoding of audio data sent in al l `RecognitionAudio` messages.", 345 "description": "*Required* Encoding of audio data sent in al l `RecognitionAudio` messages.",
329 "enum": [ 346 "enum": [
330 "ENCODING_UNSPECIFIED", 347 "ENCODING_UNSPECIFIED",
331 "LINEAR16", 348 "LINEAR16",
332 "FLAC", 349 "FLAC",
333 "MULAW", 350 "MULAW",
334 "AMR", 351 "AMR",
335 "AMR_WB" 352 "AMR_WB"
336 ], 353 ],
337 "enumDescriptions": [ 354 "enumDescriptions": [
338 "Not specified. Will return result google.rpc.Code.INVAL ID_ARGUMENT.", 355 "Not specified. Will return result google.rpc.Code.INVAL ID_ARGUMENT.",
339 "Uncompressed 16-bit signed little-endian samples (Linea r PCM).\nThis is the only encoding that may be used by `AsyncRecognize`.", 356 "Uncompressed 16-bit signed little-endian samples (Linea r PCM).\nThis is the only encoding that may be used by `AsyncRecognize`.",
340 "This is the recommended encoding for `SyncRecognize` an d\n`StreamingRecognize` because it uses lossless compression; therefore\nrecogni tion accuracy is not compromised by a lossy codec.\n\nThe stream FLAC (Free Loss less Audio Codec) encoding is specified at:\nhttp://flac.sourceforge.net/documen tation.html.\n16-bit and 24-bit samples are supported.\nNot all fields in STREAM INFO are supported.", 357 "This is the recommended encoding for `SyncRecognize` an d\n`StreamingRecognize` because it uses lossless compression; therefore\nrecogni tion accuracy is not compromised by a lossy codec.\n\nThe stream FLAC (Free Loss less Audio Codec) encoding is specified at:\nhttp://flac.sourceforge.net/documen tation.html.\n16-bit and 24-bit samples are supported.\nNot all fields in STREAM INFO are supported.",
341 "8-bit samples that compand 14-bit audio samples using G .711 PCMU/mu-law.", 358 "8-bit samples that compand 14-bit audio samples using G .711 PCMU/mu-law.",
342 "Adaptive Multi-Rate Narrowband codec. `sample_rate` mus t be 8000 Hz.", 359 "Adaptive Multi-Rate Narrowband codec. `sample_rate` mus t be 8000 Hz.",
343 "Adaptive Multi-Rate Wideband codec. `sample_rate` must be 16000 Hz." 360 "Adaptive Multi-Rate Wideband codec. `sample_rate` must be 16000 Hz."
344 ], 361 ],
345 "type": "string" 362 "type": "string"
346 }, 363 },
347 "profanityFilter": { 364 "profanityFilter": {
348 "description": "*Optional* If set to `true`, the server will attempt to filter out\nprofanities, replacing all but the initial character in each filtered word\nwith asterisks, e.g. \"f***\". If set to `false` or omitted, profanities\nwon't be filtered out.", 365 "description": "*Optional* If set to `true`, the server will attempt to filter out\nprofanities, replacing all but the initial character in each filtered word\nwith asterisks, e.g. \"f***\". If set to `false` or omitted, profanities\nwon't be filtered out.",
349 "type": "boolean" 366 "type": "boolean"
367 },
368 "maxAlternatives": {
369 "description": "*Optional* Maximum number of recognition hyp otheses to be returned.\nSpecifically, the maximum number of `SpeechRecognitionA lternative` messages\nwithin each `SpeechRecognitionResult`.\nThe server may ret urn fewer than `max_alternatives`.\nValid values are `0`-`30`. A value of `0` or `1` will return a maximum of\none. If omitted, will return a maximum of one.",
370 "format": "int32",
371 "type": "integer"
372 },
373 "sampleRate": {
374 "description": "*Required* Sample rate in Hertz of the audio data sent in all\n`RecognitionAudio` messages. Valid values are: 8000-48000.\n1 6000 is optimal. For best results, set the sampling rate of the audio\nsource to 16000 Hz. If that's not possible, use the native sample rate of\nthe audio sour ce (instead of re-sampling).",
375 "format": "int32",
376 "type": "integer"
377 },
378 "languageCode": {
379 "description": "*Optional* The language of the supplied audi o as a BCP-47 language tag.\nExample: \"en-GB\" https://www.rfc-editor.org/rfc/ bcp/bcp47.txt\nIf omitted, defaults to \"en-US\". See\n[Language Support](https: //cloud.google.com/speech/docs/languages)\nfor a list of the currently supported language codes.",
380 "type": "string"
350 } 381 }
351 }, 382 },
352 "type": "object" 383 "type": "object"
353 }, 384 },
354 "SyncRecognizeRequest": { 385 "SyncRecognizeRequest": {
355 "description": "The top-level message sent by the client for the `Sy ncRecognize` method.", 386 "description": "The top-level message sent by the client for the `Sy ncRecognize` method.",
356 "id": "SyncRecognizeRequest", 387 "id": "SyncRecognizeRequest",
357 "properties": { 388 "properties": {
358 "config": { 389 "config": {
359 "$ref": "RecognitionConfig", 390 "$ref": "RecognitionConfig",
(...skipping 80 matching lines...) Expand 10 before | Expand all | Expand 10 after
440 }, 471 },
441 "type": "array" 472 "type": "array"
442 } 473 }
443 }, 474 },
444 "type": "object" 475 "type": "object"
445 }, 476 },
446 "ListOperationsResponse": { 477 "ListOperationsResponse": {
447 "description": "The response message for Operations.ListOperations." , 478 "description": "The response message for Operations.ListOperations." ,
448 "id": "ListOperationsResponse", 479 "id": "ListOperationsResponse",
449 "properties": { 480 "properties": {
450 "nextPageToken": {
451 "description": "The standard List next-page token.",
452 "type": "string"
453 },
454 "operations": { 481 "operations": {
455 "description": "A list of operations that matches the specif ied filter in the request.", 482 "description": "A list of operations that matches the specif ied filter in the request.",
456 "items": { 483 "items": {
457 "$ref": "Operation" 484 "$ref": "Operation"
458 }, 485 },
459 "type": "array" 486 "type": "array"
487 },
488 "nextPageToken": {
489 "description": "The standard List next-page token.",
490 "type": "string"
460 } 491 }
461 }, 492 },
462 "type": "object" 493 "type": "object"
463 }, 494 },
464 "SpeechRecognitionResult": { 495 "SpeechRecognitionResult": {
465 "description": "A speech recognition result corresponding to a porti on of the audio.", 496 "description": "A speech recognition result corresponding to a porti on of the audio.",
466 "id": "SpeechRecognitionResult", 497 "id": "SpeechRecognitionResult",
467 "properties": { 498 "properties": {
468 "alternatives": { 499 "alternatives": {
469 "description": "*Output-only* May contain one or more recogn ition hypotheses (up to the\nmaximum specified in `max_alternatives`).", 500 "description": "*Output-only* May contain one or more recogn ition hypotheses (up to the\nmaximum specified in `max_alternatives`).",
470 "items": { 501 "items": {
471 "$ref": "SpeechRecognitionAlternative" 502 "$ref": "SpeechRecognitionAlternative"
472 }, 503 },
473 "type": "array" 504 "type": "array"
474 } 505 }
475 }, 506 },
476 "type": "object" 507 "type": "object"
477 },
478 "AsyncRecognizeRequest": {
479 "description": "The top-level message sent by the client for the `As yncRecognize` method.",
480 "id": "AsyncRecognizeRequest",
481 "properties": {
482 "config": {
483 "$ref": "RecognitionConfig",
484 "description": "*Required* Provides information to the recog nizer that specifies how to\nprocess the request."
485 },
486 "audio": {
487 "$ref": "RecognitionAudio",
488 "description": "*Required* The audio data to be recognized."
489 }
490 },
491 "type": "object"
492 },
493 "RecognitionAudio": {
494 "description": "Contains audio data in the encoding specified in the `RecognitionConfig`.\nEither `content` or `uri` must be supplied. Supplying bot h or neither\nreturns google.rpc.Code.INVALID_ARGUMENT. See\n[audio limits](http s://cloud.google.com/speech/limits#content).",
495 "id": "RecognitionAudio",
496 "properties": {
497 "content": {
498 "description": "The audio data bytes encoded as specified in \n`RecognitionConfig`. Note: as with all bytes fields, protobuffers use a\npure binary representation, whereas JSON representations use base64.",
499 "format": "byte",
500 "type": "string"
501 },
502 "uri": {
503 "description": "URI that points to a file that contains audi o data bytes as specified in\n`RecognitionConfig`. Currently, only Google Cloud Storage URIs are\nsupported, which must be specified in the following format:\n` gs://bucket_name/object_name` (other URI formats return\ngoogle.rpc.Code.INVALID _ARGUMENT). For more information, see\n[Request URIs](https://cloud.google.com/s torage/docs/reference-uris).",
504 "type": "string"
505 }
506 },
507 "type": "object"
508 } 508 }
509 }, 509 },
510 "servicePath": "", 510 "servicePath": "",
511 "title": "Google Cloud Speech API", 511 "title": "Google Cloud Speech API",
512 "version": "v1beta1" 512 "version": "v1beta1"
513 } 513 }
OLDNEW
« no previous file with comments | « discovery/googleapis_beta/runtimeconfig__v1beta1.json ('k') | discovery/googleapis_beta/toolresults__v1beta3.json » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698