OLD | NEW |
1 { | 1 { |
2 "auth": { | 2 "auth": { |
3 "oauth2": { | 3 "oauth2": { |
4 "scopes": { | 4 "scopes": { |
5 "https://www.googleapis.com/auth/cloud-platform": { | 5 "https://www.googleapis.com/auth/cloud-platform": { |
6 "description": "View and manage your data across Google Clou
d Platform services" | 6 "description": "View and manage your data across Google Clou
d Platform services" |
7 } | 7 } |
8 } | 8 } |
9 } | 9 } |
10 }, | 10 }, |
11 "basePath": "", | 11 "basePath": "", |
12 "baseUrl": "https://speech.googleapis.com/", | 12 "baseUrl": "https://speech.googleapis.com/", |
13 "batchPath": "batch", | 13 "batchPath": "batch", |
14 "description": "Google Cloud Speech API.", | 14 "description": "Google Cloud Speech API.", |
15 "discoveryVersion": "v1", | 15 "discoveryVersion": "v1", |
16 "documentationLink": "https://cloud.google.com/speech/", | 16 "documentationLink": "https://cloud.google.com/speech/", |
17 "icons": { | 17 "icons": { |
18 "x16": "http://www.google.com/images/icons/product/search-16.gif", | 18 "x16": "http://www.google.com/images/icons/product/search-16.gif", |
19 "x32": "http://www.google.com/images/icons/product/search-32.gif" | 19 "x32": "http://www.google.com/images/icons/product/search-32.gif" |
20 }, | 20 }, |
21 "id": "speech:v1beta1", | 21 "id": "speech:v1beta1", |
22 "kind": "discovery#restDescription", | 22 "kind": "discovery#restDescription", |
23 "name": "speech", | 23 "name": "speech", |
24 "ownerDomain": "google.com", | 24 "ownerDomain": "google.com", |
25 "ownerName": "Google", | 25 "ownerName": "Google", |
26 "parameters": { | 26 "parameters": { |
| 27 "oauth_token": { |
| 28 "description": "OAuth 2.0 token for the current user.", |
| 29 "location": "query", |
| 30 "type": "string" |
| 31 }, |
| 32 "bearer_token": { |
| 33 "description": "OAuth bearer token.", |
| 34 "location": "query", |
| 35 "type": "string" |
| 36 }, |
27 "upload_protocol": { | 37 "upload_protocol": { |
28 "description": "Upload protocol for media (e.g. \"raw\", \"multipart
\").", | 38 "description": "Upload protocol for media (e.g. \"raw\", \"multipart
\").", |
29 "location": "query", | 39 "location": "query", |
30 "type": "string" | 40 "type": "string" |
31 }, | 41 }, |
32 "prettyPrint": { | 42 "prettyPrint": { |
33 "default": "true", | 43 "default": "true", |
34 "description": "Returns response with indentations and line breaks."
, | 44 "description": "Returns response with indentations and line breaks."
, |
35 "location": "query", | 45 "location": "query", |
36 "type": "boolean" | 46 "type": "boolean" |
37 }, | 47 }, |
| 48 "fields": { |
| 49 "description": "Selector specifying which fields to include in a par
tial response.", |
| 50 "location": "query", |
| 51 "type": "string" |
| 52 }, |
38 "uploadType": { | 53 "uploadType": { |
39 "description": "Legacy upload protocol for media (e.g. \"media\", \"
multipart\").", | 54 "description": "Legacy upload protocol for media (e.g. \"media\", \"
multipart\").", |
40 "location": "query", | 55 "location": "query", |
41 "type": "string" | 56 "type": "string" |
42 }, | 57 }, |
43 "fields": { | |
44 "description": "Selector specifying which fields to include in a par
tial response.", | |
45 "location": "query", | |
46 "type": "string" | |
47 }, | |
48 "callback": { | |
49 "description": "JSONP", | |
50 "location": "query", | |
51 "type": "string" | |
52 }, | |
53 "$.xgafv": { | 58 "$.xgafv": { |
54 "description": "V1 error format.", | 59 "description": "V1 error format.", |
55 "enum": [ | 60 "enum": [ |
56 "1", | 61 "1", |
57 "2" | 62 "2" |
58 ], | 63 ], |
59 "enumDescriptions": [ | 64 "enumDescriptions": [ |
60 "v1 error format", | 65 "v1 error format", |
61 "v2 error format" | 66 "v2 error format" |
62 ], | 67 ], |
63 "location": "query", | 68 "location": "query", |
64 "type": "string" | 69 "type": "string" |
65 }, | 70 }, |
| 71 "callback": { |
| 72 "description": "JSONP", |
| 73 "location": "query", |
| 74 "type": "string" |
| 75 }, |
66 "alt": { | 76 "alt": { |
67 "default": "json", | 77 "default": "json", |
68 "description": "Data format for response.", | 78 "description": "Data format for response.", |
69 "enum": [ | 79 "enum": [ |
70 "json", | 80 "json", |
71 "media", | 81 "media", |
72 "proto" | 82 "proto" |
73 ], | 83 ], |
74 "enumDescriptions": [ | 84 "enumDescriptions": [ |
75 "Responses with Content-Type of application/json", | 85 "Responses with Content-Type of application/json", |
76 "Media download with context-dependent Content-Type", | 86 "Media download with context-dependent Content-Type", |
77 "Responses with Content-Type of application/x-protobuf" | 87 "Responses with Content-Type of application/x-protobuf" |
78 ], | 88 ], |
79 "location": "query", | 89 "location": "query", |
80 "type": "string" | 90 "type": "string" |
81 }, | 91 }, |
| 92 "access_token": { |
| 93 "description": "OAuth access token.", |
| 94 "location": "query", |
| 95 "type": "string" |
| 96 }, |
82 "key": { | 97 "key": { |
83 "description": "API key. Your API key identifies your project and pr
ovides you with API access, quota, and reports. Required unless you provide an O
Auth 2.0 token.", | 98 "description": "API key. Your API key identifies your project and pr
ovides you with API access, quota, and reports. Required unless you provide an O
Auth 2.0 token.", |
84 "location": "query", | 99 "location": "query", |
85 "type": "string" | 100 "type": "string" |
86 }, | 101 }, |
87 "access_token": { | |
88 "description": "OAuth access token.", | |
89 "location": "query", | |
90 "type": "string" | |
91 }, | |
92 "quotaUser": { | 102 "quotaUser": { |
93 "description": "Available to use for quota purposes for server-side
applications. Can be any arbitrary string assigned to a user, but should not exc
eed 40 characters.", | 103 "description": "Available to use for quota purposes for server-side
applications. Can be any arbitrary string assigned to a user, but should not exc
eed 40 characters.", |
94 "location": "query", | 104 "location": "query", |
95 "type": "string" | 105 "type": "string" |
96 }, | 106 }, |
97 "pp": { | 107 "pp": { |
98 "default": "true", | 108 "default": "true", |
99 "description": "Pretty-print response.", | 109 "description": "Pretty-print response.", |
100 "location": "query", | 110 "location": "query", |
101 "type": "boolean" | 111 "type": "boolean" |
102 }, | |
103 "bearer_token": { | |
104 "description": "OAuth bearer token.", | |
105 "location": "query", | |
106 "type": "string" | |
107 }, | |
108 "oauth_token": { | |
109 "description": "OAuth 2.0 token for the current user.", | |
110 "location": "query", | |
111 "type": "string" | |
112 } | 112 } |
113 }, | 113 }, |
114 "protocol": "rest", | 114 "protocol": "rest", |
115 "resources": { | 115 "resources": { |
116 "operations": { | 116 "operations": { |
117 "methods": { | 117 "methods": { |
118 "cancel": { | 118 "cancel": { |
119 "description": "Starts asynchronous cancellation on a long-r
unning operation. The server\nmakes a best effort to cancel the operation, but
success is not\nguaranteed. If the server doesn't support this method, it retur
ns\n`google.rpc.Code.UNIMPLEMENTED`. Clients can use\nOperations.GetOperation o
r\nother methods to check whether the cancellation succeeded or whether the\nope
ration completed despite cancellation. On successful cancellation,\nthe operatio
n is not deleted; instead, it becomes an operation with\nan Operation.error valu
e with a google.rpc.Status.code of 1,\ncorresponding to `Code.CANCELLED`.", | 119 "description": "Starts asynchronous cancellation on a long-r
unning operation. The server\nmakes a best effort to cancel the operation, but
success is not\nguaranteed. If the server doesn't support this method, it retur
ns\n`google.rpc.Code.UNIMPLEMENTED`. Clients can use\nOperations.GetOperation o
r\nother methods to check whether the cancellation succeeded or whether the\nope
ration completed despite cancellation. On successful cancellation,\nthe operatio
n is not deleted; instead, it becomes an operation with\nan Operation.error valu
e with a google.rpc.Status.code of 1,\ncorresponding to `Code.CANCELLED`.", |
120 "httpMethod": "POST", | 120 "httpMethod": "POST", |
121 "id": "speech.operations.cancel", | 121 "id": "speech.operations.cancel", |
(...skipping 40 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
162 "scopes": [ | 162 "scopes": [ |
163 "https://www.googleapis.com/auth/cloud-platform" | 163 "https://www.googleapis.com/auth/cloud-platform" |
164 ] | 164 ] |
165 }, | 165 }, |
166 "list": { | 166 "list": { |
167 "description": "Lists operations that match the specified fi
lter in the request. If the\nserver doesn't support this method, it returns `UNI
MPLEMENTED`.\n\nNOTE: the `name` binding below allows API services to override t
he binding\nto use different resource name schemes, such as `users/*/operations`
.", | 167 "description": "Lists operations that match the specified fi
lter in the request. If the\nserver doesn't support this method, it returns `UNI
MPLEMENTED`.\n\nNOTE: the `name` binding below allows API services to override t
he binding\nto use different resource name schemes, such as `users/*/operations`
.", |
168 "httpMethod": "GET", | 168 "httpMethod": "GET", |
169 "id": "speech.operations.list", | 169 "id": "speech.operations.list", |
170 "parameterOrder": [], | 170 "parameterOrder": [], |
171 "parameters": { | 171 "parameters": { |
172 "filter": { | |
173 "description": "The standard list filter.", | |
174 "location": "query", | |
175 "type": "string" | |
176 }, | |
177 "name": { | 172 "name": { |
178 "description": "The name of the operation collection
.", | 173 "description": "The name of the operation collection
.", |
179 "location": "query", | 174 "location": "query", |
180 "type": "string" | 175 "type": "string" |
181 }, | 176 }, |
182 "pageToken": { | 177 "pageToken": { |
183 "description": "The standard list page token.", | 178 "description": "The standard list page token.", |
184 "location": "query", | 179 "location": "query", |
185 "type": "string" | 180 "type": "string" |
186 }, | 181 }, |
187 "pageSize": { | 182 "pageSize": { |
188 "description": "The standard list page size.", | 183 "description": "The standard list page size.", |
189 "format": "int32", | 184 "format": "int32", |
190 "location": "query", | 185 "location": "query", |
191 "type": "integer" | 186 "type": "integer" |
| 187 }, |
| 188 "filter": { |
| 189 "description": "The standard list filter.", |
| 190 "location": "query", |
| 191 "type": "string" |
192 } | 192 } |
193 }, | 193 }, |
194 "path": "v1beta1/operations", | 194 "path": "v1beta1/operations", |
195 "response": { | 195 "response": { |
196 "$ref": "ListOperationsResponse" | 196 "$ref": "ListOperationsResponse" |
197 }, | 197 }, |
198 "scopes": [ | 198 "scopes": [ |
199 "https://www.googleapis.com/auth/cloud-platform" | 199 "https://www.googleapis.com/auth/cloud-platform" |
200 ] | 200 ] |
201 }, | 201 }, |
(...skipping 55 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
257 "response": { | 257 "response": { |
258 "$ref": "SyncRecognizeResponse" | 258 "$ref": "SyncRecognizeResponse" |
259 }, | 259 }, |
260 "scopes": [ | 260 "scopes": [ |
261 "https://www.googleapis.com/auth/cloud-platform" | 261 "https://www.googleapis.com/auth/cloud-platform" |
262 ] | 262 ] |
263 } | 263 } |
264 } | 264 } |
265 } | 265 } |
266 }, | 266 }, |
267 "revision": "20170303", | 267 "revision": "20170323", |
268 "rootUrl": "https://speech.googleapis.com/", | 268 "rootUrl": "https://speech.googleapis.com/", |
269 "schemas": { | 269 "schemas": { |
| 270 "RecognitionAudio": { |
| 271 "description": "Contains audio data in the encoding specified in the
`RecognitionConfig`.\nEither `content` or `uri` must be supplied. Supplying bot
h or neither\nreturns google.rpc.Code.INVALID_ARGUMENT. See\n[audio limits](http
s://cloud.google.com/speech/limits#content).", |
| 272 "id": "RecognitionAudio", |
| 273 "properties": { |
| 274 "content": { |
| 275 "description": "The audio data bytes encoded as specified in
\n`RecognitionConfig`. Note: as with all bytes fields, protobuffers use a\npure
binary representation, whereas JSON representations use base64.", |
| 276 "format": "byte", |
| 277 "type": "string" |
| 278 }, |
| 279 "uri": { |
| 280 "description": "URI that points to a file that contains audi
o data bytes as specified in\n`RecognitionConfig`. Currently, only Google Cloud
Storage URIs are\nsupported, which must be specified in the following format:\n`
gs://bucket_name/object_name` (other URI formats return\ngoogle.rpc.Code.INVALID
_ARGUMENT). For more information, see\n[Request URIs](https://cloud.google.com/s
torage/docs/reference-uris).", |
| 281 "type": "string" |
| 282 } |
| 283 }, |
| 284 "type": "object" |
| 285 }, |
| 286 "AsyncRecognizeRequest": { |
| 287 "description": "The top-level message sent by the client for the `As
yncRecognize` method.", |
| 288 "id": "AsyncRecognizeRequest", |
| 289 "properties": { |
| 290 "config": { |
| 291 "$ref": "RecognitionConfig", |
| 292 "description": "*Required* Provides information to the recog
nizer that specifies how to\nprocess the request." |
| 293 }, |
| 294 "audio": { |
| 295 "$ref": "RecognitionAudio", |
| 296 "description": "*Required* The audio data to be recognized." |
| 297 } |
| 298 }, |
| 299 "type": "object" |
| 300 }, |
270 "Operation": { | 301 "Operation": { |
271 "description": "This resource represents a long-running operation th
at is the result of a\nnetwork API call.", | 302 "description": "This resource represents a long-running operation th
at is the result of a\nnetwork API call.", |
272 "id": "Operation", | 303 "id": "Operation", |
273 "properties": { | 304 "properties": { |
274 "done": { | 305 "done": { |
275 "description": "If the value is `false`, it means the operat
ion is still in progress.\nIf true, the operation is completed, and either `erro
r` or `response` is\navailable.", | 306 "description": "If the value is `false`, it means the operat
ion is still in progress.\nIf true, the operation is completed, and either `erro
r` or `response` is\navailable.", |
276 "type": "boolean" | 307 "type": "boolean" |
277 }, | 308 }, |
278 "response": { | 309 "response": { |
279 "additionalProperties": { | 310 "additionalProperties": { |
(...skipping 19 matching lines...) Expand all Loading... |
299 "description": "Service-specific metadata associated with th
e operation. It typically\ncontains progress information and common metadata su
ch as create time.\nSome services might not provide such metadata. Any method t
hat returns a\nlong-running operation should document the metadata type, if any.
", | 330 "description": "Service-specific metadata associated with th
e operation. It typically\ncontains progress information and common metadata su
ch as create time.\nSome services might not provide such metadata. Any method t
hat returns a\nlong-running operation should document the metadata type, if any.
", |
300 "type": "object" | 331 "type": "object" |
301 } | 332 } |
302 }, | 333 }, |
303 "type": "object" | 334 "type": "object" |
304 }, | 335 }, |
305 "RecognitionConfig": { | 336 "RecognitionConfig": { |
306 "description": "Provides information to the recognizer that specifie
s how to process the\nrequest.", | 337 "description": "Provides information to the recognizer that specifie
s how to process the\nrequest.", |
307 "id": "RecognitionConfig", | 338 "id": "RecognitionConfig", |
308 "properties": { | 339 "properties": { |
309 "maxAlternatives": { | |
310 "description": "*Optional* Maximum number of recognition hyp
otheses to be returned.\nSpecifically, the maximum number of `SpeechRecognitionA
lternative` messages\nwithin each `SpeechRecognitionResult`.\nThe server may ret
urn fewer than `max_alternatives`.\nValid values are `0`-`30`. A value of `0` or
`1` will return a maximum of\none. If omitted, will return a maximum of one.", | |
311 "format": "int32", | |
312 "type": "integer" | |
313 }, | |
314 "sampleRate": { | |
315 "description": "*Required* Sample rate in Hertz of the audio
data sent in all\n`RecognitionAudio` messages. Valid values are: 8000-48000.\n1
6000 is optimal. For best results, set the sampling rate of the audio\nsource to
16000 Hz. If that's not possible, use the native sample rate of\nthe audio sour
ce (instead of re-sampling).", | |
316 "format": "int32", | |
317 "type": "integer" | |
318 }, | |
319 "languageCode": { | |
320 "description": "*Optional* The language of the supplied audi
o as a BCP-47 language tag.\nExample: \"en-GB\" https://www.rfc-editor.org/rfc/
bcp/bcp47.txt\nIf omitted, defaults to \"en-US\". See\n[Language Support](https:
//cloud.google.com/speech/docs/languages)\nfor a list of the currently supported
language codes.", | |
321 "type": "string" | |
322 }, | |
323 "speechContext": { | 340 "speechContext": { |
324 "$ref": "SpeechContext", | 341 "$ref": "SpeechContext", |
325 "description": "*Optional* A means to provide context to ass
ist the speech recognition." | 342 "description": "*Optional* A means to provide context to ass
ist the speech recognition." |
326 }, | 343 }, |
327 "encoding": { | 344 "encoding": { |
328 "description": "*Required* Encoding of audio data sent in al
l `RecognitionAudio` messages.", | 345 "description": "*Required* Encoding of audio data sent in al
l `RecognitionAudio` messages.", |
329 "enum": [ | 346 "enum": [ |
330 "ENCODING_UNSPECIFIED", | 347 "ENCODING_UNSPECIFIED", |
331 "LINEAR16", | 348 "LINEAR16", |
332 "FLAC", | 349 "FLAC", |
333 "MULAW", | 350 "MULAW", |
334 "AMR", | 351 "AMR", |
335 "AMR_WB" | 352 "AMR_WB" |
336 ], | 353 ], |
337 "enumDescriptions": [ | 354 "enumDescriptions": [ |
338 "Not specified. Will return result google.rpc.Code.INVAL
ID_ARGUMENT.", | 355 "Not specified. Will return result google.rpc.Code.INVAL
ID_ARGUMENT.", |
339 "Uncompressed 16-bit signed little-endian samples (Linea
r PCM).\nThis is the only encoding that may be used by `AsyncRecognize`.", | 356 "Uncompressed 16-bit signed little-endian samples (Linea
r PCM).\nThis is the only encoding that may be used by `AsyncRecognize`.", |
340 "This is the recommended encoding for `SyncRecognize` an
d\n`StreamingRecognize` because it uses lossless compression; therefore\nrecogni
tion accuracy is not compromised by a lossy codec.\n\nThe stream FLAC (Free Loss
less Audio Codec) encoding is specified at:\nhttp://flac.sourceforge.net/documen
tation.html.\n16-bit and 24-bit samples are supported.\nNot all fields in STREAM
INFO are supported.", | 357 "This is the recommended encoding for `SyncRecognize` an
d\n`StreamingRecognize` because it uses lossless compression; therefore\nrecogni
tion accuracy is not compromised by a lossy codec.\n\nThe stream FLAC (Free Loss
less Audio Codec) encoding is specified at:\nhttp://flac.sourceforge.net/documen
tation.html.\n16-bit and 24-bit samples are supported.\nNot all fields in STREAM
INFO are supported.", |
341 "8-bit samples that compand 14-bit audio samples using G
.711 PCMU/mu-law.", | 358 "8-bit samples that compand 14-bit audio samples using G
.711 PCMU/mu-law.", |
342 "Adaptive Multi-Rate Narrowband codec. `sample_rate` mus
t be 8000 Hz.", | 359 "Adaptive Multi-Rate Narrowband codec. `sample_rate` mus
t be 8000 Hz.", |
343 "Adaptive Multi-Rate Wideband codec. `sample_rate` must
be 16000 Hz." | 360 "Adaptive Multi-Rate Wideband codec. `sample_rate` must
be 16000 Hz." |
344 ], | 361 ], |
345 "type": "string" | 362 "type": "string" |
346 }, | 363 }, |
347 "profanityFilter": { | 364 "profanityFilter": { |
348 "description": "*Optional* If set to `true`, the server will
attempt to filter out\nprofanities, replacing all but the initial character in
each filtered word\nwith asterisks, e.g. \"f***\". If set to `false` or omitted,
profanities\nwon't be filtered out.", | 365 "description": "*Optional* If set to `true`, the server will
attempt to filter out\nprofanities, replacing all but the initial character in
each filtered word\nwith asterisks, e.g. \"f***\". If set to `false` or omitted,
profanities\nwon't be filtered out.", |
349 "type": "boolean" | 366 "type": "boolean" |
| 367 }, |
| 368 "maxAlternatives": { |
| 369 "description": "*Optional* Maximum number of recognition hyp
otheses to be returned.\nSpecifically, the maximum number of `SpeechRecognitionA
lternative` messages\nwithin each `SpeechRecognitionResult`.\nThe server may ret
urn fewer than `max_alternatives`.\nValid values are `0`-`30`. A value of `0` or
`1` will return a maximum of\none. If omitted, will return a maximum of one.", |
| 370 "format": "int32", |
| 371 "type": "integer" |
| 372 }, |
| 373 "sampleRate": { |
| 374 "description": "*Required* Sample rate in Hertz of the audio
data sent in all\n`RecognitionAudio` messages. Valid values are: 8000-48000.\n1
6000 is optimal. For best results, set the sampling rate of the audio\nsource to
16000 Hz. If that's not possible, use the native sample rate of\nthe audio sour
ce (instead of re-sampling).", |
| 375 "format": "int32", |
| 376 "type": "integer" |
| 377 }, |
| 378 "languageCode": { |
| 379 "description": "*Optional* The language of the supplied audi
o as a BCP-47 language tag.\nExample: \"en-GB\" https://www.rfc-editor.org/rfc/
bcp/bcp47.txt\nIf omitted, defaults to \"en-US\". See\n[Language Support](https:
//cloud.google.com/speech/docs/languages)\nfor a list of the currently supported
language codes.", |
| 380 "type": "string" |
350 } | 381 } |
351 }, | 382 }, |
352 "type": "object" | 383 "type": "object" |
353 }, | 384 }, |
354 "SyncRecognizeRequest": { | 385 "SyncRecognizeRequest": { |
355 "description": "The top-level message sent by the client for the `Sy
ncRecognize` method.", | 386 "description": "The top-level message sent by the client for the `Sy
ncRecognize` method.", |
356 "id": "SyncRecognizeRequest", | 387 "id": "SyncRecognizeRequest", |
357 "properties": { | 388 "properties": { |
358 "config": { | 389 "config": { |
359 "$ref": "RecognitionConfig", | 390 "$ref": "RecognitionConfig", |
(...skipping 80 matching lines...) Expand 10 before | Expand all | Expand 10 after Loading... |
440 }, | 471 }, |
441 "type": "array" | 472 "type": "array" |
442 } | 473 } |
443 }, | 474 }, |
444 "type": "object" | 475 "type": "object" |
445 }, | 476 }, |
446 "ListOperationsResponse": { | 477 "ListOperationsResponse": { |
447 "description": "The response message for Operations.ListOperations."
, | 478 "description": "The response message for Operations.ListOperations."
, |
448 "id": "ListOperationsResponse", | 479 "id": "ListOperationsResponse", |
449 "properties": { | 480 "properties": { |
450 "nextPageToken": { | |
451 "description": "The standard List next-page token.", | |
452 "type": "string" | |
453 }, | |
454 "operations": { | 481 "operations": { |
455 "description": "A list of operations that matches the specif
ied filter in the request.", | 482 "description": "A list of operations that matches the specif
ied filter in the request.", |
456 "items": { | 483 "items": { |
457 "$ref": "Operation" | 484 "$ref": "Operation" |
458 }, | 485 }, |
459 "type": "array" | 486 "type": "array" |
| 487 }, |
| 488 "nextPageToken": { |
| 489 "description": "The standard List next-page token.", |
| 490 "type": "string" |
460 } | 491 } |
461 }, | 492 }, |
462 "type": "object" | 493 "type": "object" |
463 }, | 494 }, |
464 "SpeechRecognitionResult": { | 495 "SpeechRecognitionResult": { |
465 "description": "A speech recognition result corresponding to a porti
on of the audio.", | 496 "description": "A speech recognition result corresponding to a porti
on of the audio.", |
466 "id": "SpeechRecognitionResult", | 497 "id": "SpeechRecognitionResult", |
467 "properties": { | 498 "properties": { |
468 "alternatives": { | 499 "alternatives": { |
469 "description": "*Output-only* May contain one or more recogn
ition hypotheses (up to the\nmaximum specified in `max_alternatives`).", | 500 "description": "*Output-only* May contain one or more recogn
ition hypotheses (up to the\nmaximum specified in `max_alternatives`).", |
470 "items": { | 501 "items": { |
471 "$ref": "SpeechRecognitionAlternative" | 502 "$ref": "SpeechRecognitionAlternative" |
472 }, | 503 }, |
473 "type": "array" | 504 "type": "array" |
474 } | 505 } |
475 }, | 506 }, |
476 "type": "object" | 507 "type": "object" |
477 }, | |
478 "AsyncRecognizeRequest": { | |
479 "description": "The top-level message sent by the client for the `As
yncRecognize` method.", | |
480 "id": "AsyncRecognizeRequest", | |
481 "properties": { | |
482 "config": { | |
483 "$ref": "RecognitionConfig", | |
484 "description": "*Required* Provides information to the recog
nizer that specifies how to\nprocess the request." | |
485 }, | |
486 "audio": { | |
487 "$ref": "RecognitionAudio", | |
488 "description": "*Required* The audio data to be recognized." | |
489 } | |
490 }, | |
491 "type": "object" | |
492 }, | |
493 "RecognitionAudio": { | |
494 "description": "Contains audio data in the encoding specified in the
`RecognitionConfig`.\nEither `content` or `uri` must be supplied. Supplying bot
h or neither\nreturns google.rpc.Code.INVALID_ARGUMENT. See\n[audio limits](http
s://cloud.google.com/speech/limits#content).", | |
495 "id": "RecognitionAudio", | |
496 "properties": { | |
497 "content": { | |
498 "description": "The audio data bytes encoded as specified in
\n`RecognitionConfig`. Note: as with all bytes fields, protobuffers use a\npure
binary representation, whereas JSON representations use base64.", | |
499 "format": "byte", | |
500 "type": "string" | |
501 }, | |
502 "uri": { | |
503 "description": "URI that points to a file that contains audi
o data bytes as specified in\n`RecognitionConfig`. Currently, only Google Cloud
Storage URIs are\nsupported, which must be specified in the following format:\n`
gs://bucket_name/object_name` (other URI formats return\ngoogle.rpc.Code.INVALID
_ARGUMENT). For more information, see\n[Request URIs](https://cloud.google.com/s
torage/docs/reference-uris).", | |
504 "type": "string" | |
505 } | |
506 }, | |
507 "type": "object" | |
508 } | 508 } |
509 }, | 509 }, |
510 "servicePath": "", | 510 "servicePath": "", |
511 "title": "Google Cloud Speech API", | 511 "title": "Google Cloud Speech API", |
512 "version": "v1beta1" | 512 "version": "v1beta1" |
513 } | 513 } |
OLD | NEW |