Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(456)

Unified Diff: generated/googleapis/lib/speech/v1.dart

Issue 3006323002: Api-Roll 54: 2017-09-11 (Closed)
Patch Set: use 2.0.0-dev.infinity sdk constraint in pubspecs Created 3 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « generated/googleapis/lib/spanner/v1.dart ('k') | generated/googleapis/lib/storage/v1.dart » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: generated/googleapis/lib/speech/v1.dart
diff --git a/generated/googleapis/lib/speech/v1.dart b/generated/googleapis/lib/speech/v1.dart
index 5b0c2b802f78e0923078fcc9005e584a52dc4a43..4b2431e06bce18b0e3fc956e95f3e0ef81313cce 100644
--- a/generated/googleapis/lib/speech/v1.dart
+++ b/generated/googleapis/lib/speech/v1.dart
@@ -9,60 +9,59 @@ import 'dart:convert' as convert;
import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons;
import 'package:http/http.dart' as http;
-export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show
- ApiRequestError, DetailedApiRequestError;
+export 'package:_discoveryapis_commons/_discoveryapis_commons.dart'
+ show ApiRequestError, DetailedApiRequestError;
const core.String USER_AGENT = 'dart-api-client speech/v1';
-/** Converts audio to text by applying powerful neural network models. */
+/// Converts audio to text by applying powerful neural network models.
class SpeechApi {
- /** View and manage your data across Google Cloud Platform services */
- static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platform";
-
+ /// View and manage your data across Google Cloud Platform services
+ static const CloudPlatformScope =
+ "https://www.googleapis.com/auth/cloud-platform";
final commons.ApiRequester _requester;
OperationsResourceApi get operations => new OperationsResourceApi(_requester);
SpeechResourceApi get speech => new SpeechResourceApi(_requester);
- SpeechApi(http.Client client, {core.String rootUrl: "https://speech.googleapis.com/", core.String servicePath: ""}) :
- _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
+ SpeechApi(http.Client client,
+ {core.String rootUrl: "https://speech.googleapis.com/",
+ core.String servicePath: ""})
+ : _requester =
+ new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
}
-
class OperationsResourceApi {
final commons.ApiRequester _requester;
- OperationsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Starts asynchronous cancellation on a long-running operation. The server
- * makes a best effort to cancel the operation, but success is not
- * guaranteed. If the server doesn't support this method, it returns
- * `google.rpc.Code.UNIMPLEMENTED`. Clients can use
- * Operations.GetOperation or
- * other methods to check whether the cancellation succeeded or whether the
- * operation completed despite cancellation. On successful cancellation,
- * the operation is not deleted; instead, it becomes an operation with
- * an Operation.error value with a google.rpc.Status.code of 1,
- * corresponding to `Code.CANCELLED`.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * [name] - The name of the operation resource to be cancelled.
- * Value must have pattern "^[^/]+$".
- *
- * Completes with a [Empty].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
+ OperationsResourceApi(commons.ApiRequester client) : _requester = client;
+
+ /// Starts asynchronous cancellation on a long-running operation. The server
+ /// makes a best effort to cancel the operation, but success is not
+ /// guaranteed. If the server doesn't support this method, it returns
+ /// `google.rpc.Code.UNIMPLEMENTED`. Clients can use
+ /// Operations.GetOperation or
+ /// other methods to check whether the cancellation succeeded or whether the
+ /// operation completed despite cancellation. On successful cancellation,
+ /// the operation is not deleted; instead, it becomes an operation with
+ /// an Operation.error value with a google.rpc.Status.code of 1,
+ /// corresponding to `Code.CANCELLED`.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// [name] - The name of the operation resource to be cancelled.
+ /// Value must have pattern "^[^/]+$".
+ ///
+ /// Completes with a [Empty].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
async.Future<Empty> cancel(CancelOperationRequest request, core.String name) {
var _url = null;
var _queryParams = new core.Map();
@@ -78,37 +77,37 @@ class OperationsResourceApi {
throw new core.ArgumentError("Parameter name is required.");
}
- _url = 'v1/operations/' + commons.Escaper.ecapeVariableReserved('$name') + ':cancel';
+ _url = 'v1/operations/' +
+ commons.Escaper.ecapeVariableReserved('$name') +
+ ':cancel';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Empty.fromJson(data));
}
- /**
- * Deletes a long-running operation. This method indicates that the client is
- * no longer interested in the operation result. It does not cancel the
- * operation. If the server doesn't support this method, it returns
- * `google.rpc.Code.UNIMPLEMENTED`.
- *
- * Request parameters:
- *
- * [name] - The name of the operation resource to be deleted.
- * Value must have pattern "^[^/]+$".
- *
- * Completes with a [Empty].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
+ /// Deletes a long-running operation. This method indicates that the client
+ /// is
+ /// no longer interested in the operation result. It does not cancel the
+ /// operation. If the server doesn't support this method, it returns
+ /// `google.rpc.Code.UNIMPLEMENTED`.
+ ///
+ /// Request parameters:
+ ///
+ /// [name] - The name of the operation resource to be deleted.
+ /// Value must have pattern "^[^/]+$".
+ ///
+ /// Completes with a [Empty].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
async.Future<Empty> delete(core.String name) {
var _url = null;
var _queryParams = new core.Map();
@@ -123,34 +122,32 @@ class OperationsResourceApi {
_url = 'v1/operations/' + commons.Escaper.ecapeVariableReserved('$name');
- var _response = _requester.request(_url,
- "DELETE",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "DELETE",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Empty.fromJson(data));
}
- /**
- * Gets the latest state of a long-running operation. Clients can use this
- * method to poll the operation result at intervals as recommended by the API
- * service.
- *
- * Request parameters:
- *
- * [name] - The name of the operation resource.
- * Value must have pattern "^[^/]+$".
- *
- * Completes with a [Operation].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
+ /// Gets the latest state of a long-running operation. Clients can use this
+ /// method to poll the operation result at intervals as recommended by the
+ /// API
+ /// service.
+ ///
+ /// Request parameters:
+ ///
+ /// [name] - The name of the operation resource.
+ /// Value must have pattern "^[^/]+$".
+ ///
+ /// Completes with a [Operation].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
async.Future<Operation> get(core.String name) {
var _url = null;
var _queryParams = new core.Map();
@@ -165,47 +162,49 @@ class OperationsResourceApi {
_url = 'v1/operations/' + commons.Escaper.ecapeVariableReserved('$name');
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Operation.fromJson(data));
}
- /**
- * Lists operations that match the specified filter in the request. If the
- * server doesn't support this method, it returns `UNIMPLEMENTED`.
- *
- * NOTE: the `name` binding allows API services to override the binding
- * to use different resource name schemes, such as `users / * /operations`. To
- * override the binding, API services can add a binding such as
- * `"/v1/{name=users / * }/operations"` to their service configuration.
- * For backwards compatibility, the default name includes the operations
- * collection id, however overriding users must ensure the name binding
- * is the parent resource, without the operations collection id.
- *
- * Request parameters:
- *
- * [filter] - The standard list filter.
- *
- * [pageToken] - The standard list page token.
- *
- * [name] - The name of the operation's parent resource.
- *
- * [pageSize] - The standard list page size.
- *
- * Completes with a [ListOperationsResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<ListOperationsResponse> list({core.String filter, core.String pageToken, core.String name, core.int pageSize}) {
+ /// Lists operations that match the specified filter in the request. If the
+ /// server doesn't support this method, it returns `UNIMPLEMENTED`.
+ ///
+ /// NOTE: the `name` binding allows API services to override the binding
+ /// to use different resource name schemes, such as `users / * /operations`.
+ /// To
+ /// override the binding, API services can add a binding such as
+ /// `"/v1/{name=users / * }/operations"` to their service configuration.
+ /// For backwards compatibility, the default name includes the operations
+ /// collection id, however overriding users must ensure the name binding
+ /// is the parent resource, without the operations collection id.
+ ///
+ /// Request parameters:
+ ///
+ /// [filter] - The standard list filter.
+ ///
+ /// [pageToken] - The standard list page token.
+ ///
+ /// [name] - The name of the operation's parent resource.
+ ///
+ /// [pageSize] - The standard list page size.
+ ///
+ /// Completes with a [ListOperationsResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<ListOperationsResponse> list(
+ {core.String filter,
+ core.String pageToken,
+ core.String name,
+ core.int pageSize}) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -228,44 +227,39 @@ class OperationsResourceApi {
_url = 'v1/operations';
- var _response = _requester.request(_url,
- "GET",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "GET",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new ListOperationsResponse.fromJson(data));
}
-
}
-
class SpeechResourceApi {
final commons.ApiRequester _requester;
- SpeechResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Performs asynchronous speech recognition: receive results via the
- * google.longrunning.Operations interface. Returns either an
- * `Operation.error` or an `Operation.response` which contains
- * a `LongRunningRecognizeResponse` message.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [Operation].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<Operation> longrunningrecognize(LongRunningRecognizeRequest request) {
+ SpeechResourceApi(commons.ApiRequester client) : _requester = client;
+
+ /// Performs asynchronous speech recognition: receive results via the
+ /// google.longrunning.Operations interface. Returns either an
+ /// `Operation.error` or an `Operation.response` which contains
+ /// a `LongRunningRecognizeResponse` message.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [Operation].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<Operation> longrunningrecognize(
+ LongRunningRecognizeRequest request) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -279,32 +273,29 @@ class SpeechResourceApi {
_url = 'v1/speech:longrunningrecognize';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new Operation.fromJson(data));
}
- /**
- * Performs synchronous speech recognition: receive results after all audio
- * has been sent and processed.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [RecognizeResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
+ /// Performs synchronous speech recognition: receive results after all audio
+ /// has been sent and processed.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [RecognizeResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
async.Future<RecognizeResponse> recognize(RecognizeRequest request) {
var _url = null;
var _queryParams = new core.Map();
@@ -319,63 +310,56 @@ class SpeechResourceApi {
_url = 'v1/speech:recognize';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new RecognizeResponse.fromJson(data));
}
-
}
-
-
-/** The request message for Operations.CancelOperation. */
+/// The request message for Operations.CancelOperation.
class CancelOperationRequest {
-
CancelOperationRequest();
- CancelOperationRequest.fromJson(core.Map _json) {
- }
+ CancelOperationRequest.fromJson(core.Map _json) {}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
return _json;
}
}
-/**
- * A generic empty message that you can re-use to avoid defining duplicated
- * empty messages in your APIs. A typical example is to use it as the request
- * or the response type of an API method. For instance:
- *
- * service Foo {
- * rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
- * }
- *
- * The JSON representation for `Empty` is empty JSON object `{}`.
- */
+/// A generic empty message that you can re-use to avoid defining duplicated
+/// empty messages in your APIs. A typical example is to use it as the request
+/// or the response type of an API method. For instance:
+///
+/// service Foo {
+/// rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty);
+/// }
+///
+/// The JSON representation for `Empty` is empty JSON object `{}`.
class Empty {
-
Empty();
- Empty.fromJson(core.Map _json) {
- }
+ Empty.fromJson(core.Map _json) {}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
return _json;
}
}
-/** The response message for Operations.ListOperations. */
+/// The response message for Operations.ListOperations.
class ListOperationsResponse {
- /** The standard List next-page token. */
+ /// The standard List next-page token.
core.String nextPageToken;
- /** A list of operations that matches the specified filter in the request. */
+
+ /// A list of operations that matches the specified filter in the request.
core.List<Operation> operations;
ListOperationsResponse();
@@ -385,33 +369,34 @@ class ListOperationsResponse {
nextPageToken = _json["nextPageToken"];
}
if (_json.containsKey("operations")) {
- operations = _json["operations"].map((value) => new Operation.fromJson(value)).toList();
+ operations = _json["operations"]
+ .map((value) => new Operation.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (nextPageToken != null) {
_json["nextPageToken"] = nextPageToken;
}
if (operations != null) {
- _json["operations"] = operations.map((value) => (value).toJson()).toList();
+ _json["operations"] =
+ operations.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * The top-level message sent by the client for the `LongRunningRecognize`
- * method.
- */
+/// The top-level message sent by the client for the `LongRunningRecognize`
+/// method.
class LongRunningRecognizeRequest {
- /** *Required* The audio data to be recognized. */
+ /// *Required* The audio data to be recognized.
RecognitionAudio audio;
- /**
- * *Required* Provides information to the recognizer that specifies how to
- * process the request.
- */
+
+ /// *Required* Provides information to the recognizer that specifies how to
+ /// process the request.
RecognitionConfig config;
LongRunningRecognizeRequest();
@@ -426,7 +411,8 @@ class LongRunningRecognizeRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (audio != null) {
_json["audio"] = (audio).toJson();
}
@@ -437,48 +423,44 @@ class LongRunningRecognizeRequest {
}
}
-/**
- * This resource represents a long-running operation that is the result of a
- * network API call.
- */
+/// This resource represents a long-running operation that is the result of a
+/// network API call.
class Operation {
- /**
- * If the value is `false`, it means the operation is still in progress.
- * If true, the operation is completed, and either `error` or `response` is
- * available.
- */
+ /// If the value is `false`, it means the operation is still in progress.
+ /// If `true`, the operation is completed, and either `error` or `response`
+ /// is
+ /// available.
core.bool done;
- /** The error result of the operation in case of failure or cancellation. */
+
+ /// The error result of the operation in case of failure or cancellation.
Status error;
- /**
- * Service-specific metadata associated with the operation. It typically
- * contains progress information and common metadata such as create time.
- * Some services might not provide such metadata. Any method that returns a
- * long-running operation should document the metadata type, if any.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// Service-specific metadata associated with the operation. It typically
+ /// contains progress information and common metadata such as create time.
+ /// Some services might not provide such metadata. Any method that returns a
+ /// long-running operation should document the metadata type, if any.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> metadata;
- /**
- * The server-assigned name, which is only unique within the same service that
- * originally returns it. If you use the default HTTP mapping, the
- * `name` should have the format of `operations/some/unique/name`.
- */
+
+ /// The server-assigned name, which is only unique within the same service
+ /// that
+ /// originally returns it. If you use the default HTTP mapping, the
+ /// `name` should have the format of `operations/some/unique/name`.
core.String name;
- /**
- * The normal response of the operation in case of success. If the original
- * method returns no data on success, such as `Delete`, the response is
- * `google.protobuf.Empty`. If the original method is standard
- * `Get`/`Create`/`Update`, the response should be the resource. For other
- * methods, the response should have the type `XxxResponse`, where `Xxx`
- * is the original method name. For example, if the original method name
- * is `TakeSnapshot()`, the inferred response type is
- * `TakeSnapshotResponse`.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// The normal response of the operation in case of success. If the original
+ /// method returns no data on success, such as `Delete`, the response is
+ /// `google.protobuf.Empty`. If the original method is standard
+ /// `Get`/`Create`/`Update`, the response should be the resource. For other
+ /// methods, the response should have the type `XxxResponse`, where `Xxx`
+ /// is the original method name. For example, if the original method name
+ /// is `TakeSnapshot()`, the inferred response type is
+ /// `TakeSnapshotResponse`.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.Map<core.String, core.Object> response;
Operation();
@@ -502,7 +484,8 @@ class Operation {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (done != null) {
_json["done"] = done;
}
@@ -522,34 +505,30 @@ class Operation {
}
}
-/**
- * Contains audio data in the encoding specified in the `RecognitionConfig`.
- * Either `content` or `uri` must be supplied. Supplying both or neither
- * returns google.rpc.Code.INVALID_ARGUMENT. See
- * [audio limits](https://cloud.google.com/speech/limits#content).
- */
+/// Contains audio data in the encoding specified in the `RecognitionConfig`.
+/// Either `content` or `uri` must be supplied. Supplying both or neither
+/// returns google.rpc.Code.INVALID_ARGUMENT. See
+/// [audio limits](https://cloud.google.com/speech/limits#content).
class RecognitionAudio {
- /**
- * The audio data bytes encoded as specified in
- * `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
- * pure binary representation, whereas JSON representations use base64.
- */
+ /// The audio data bytes encoded as specified in
+ /// `RecognitionConfig`. Note: as with all bytes fields, protobuffers use a
+ /// pure binary representation, whereas JSON representations use base64.
core.String content;
core.List<core.int> get contentAsBytes {
return convert.BASE64.decode(content);
}
void set contentAsBytes(core.List<core.int> _bytes) {
- content = convert.BASE64.encode(_bytes).replaceAll("/", "_").replaceAll("+", "-");
+ content =
+ convert.BASE64.encode(_bytes).replaceAll("/", "_").replaceAll("+", "-");
}
- /**
- * URI that points to a file that contains audio data bytes as specified in
- * `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
- * supported, which must be specified in the following format:
- * `gs://bucket_name/object_name` (other URI formats return
- * google.rpc.Code.INVALID_ARGUMENT). For more information, see
- * [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
- */
+
+ /// URI that points to a file that contains audio data bytes as specified in
+ /// `RecognitionConfig`. Currently, only Google Cloud Storage URIs are
+ /// supported, which must be specified in the following format:
+ /// `gs://bucket_name/object_name` (other URI formats return
+ /// google.rpc.Code.INVALID_ARGUMENT). For more information, see
+ /// [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
core.String uri;
RecognitionAudio();
@@ -564,7 +543,8 @@ class RecognitionAudio {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (content != null) {
_json["content"] = content;
}
@@ -575,92 +555,86 @@ class RecognitionAudio {
}
}
-/**
- * Provides information to the recognizer that specifies how to process the
- * request.
- */
+/// Provides information to the recognizer that specifies how to process the
+/// request.
class RecognitionConfig {
- /**
- * *Optional* If `true`, the top result includes a list of words and
- * the start and end time offsets (timestamps) for those words. If
- * `false`, no word-level time offset information is returned. The default is
- * `false`.
- */
+ /// *Optional* If `true`, the top result includes a list of words and
+ /// the start and end time offsets (timestamps) for those words. If
+ /// `false`, no word-level time offset information is returned. The default
+ /// is
+ /// `false`.
core.bool enableWordTimeOffsets;
- /**
- * *Required* Encoding of audio data sent in all `RecognitionAudio` messages.
- * Possible string values are:
- * - "ENCODING_UNSPECIFIED" : Not specified. Will return result
- * google.rpc.Code.INVALID_ARGUMENT.
- * - "LINEAR16" : Uncompressed 16-bit signed little-endian samples (Linear
- * PCM).
- * - "FLAC" : [`FLAC`](https://xiph.org/flac/documentation.html) (Free
- * Lossless Audio
- * Codec) is the recommended encoding because it is
- * lossless--therefore recognition is not compromised--and
- * requires only about half the bandwidth of `LINEAR16`. `FLAC` stream
- * encoding supports 16-bit and 24-bit samples, however, not all fields in
- * `STREAMINFO` are supported.
- * - "MULAW" : 8-bit samples that compand 14-bit audio samples using G.711
- * PCMU/mu-law.
- * - "AMR" : Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must be
- * 8000.
- * - "AMR_WB" : Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must
- * be 16000.
- * - "OGG_OPUS" : Opus encoded audio frames in Ogg container
- * ([OggOpus](https://wiki.xiph.org/OggOpus)).
- * `sample_rate_hertz` must be 16000.
- * - "SPEEX_WITH_HEADER_BYTE" : Although the use of lossy encodings is not
- * recommended, if a very low
- * bitrate encoding is required, `OGG_OPUS` is highly preferred over
- * Speex encoding. The [Speex](https://speex.org/) encoding supported by
- * Cloud Speech API has a header byte in each block, as in MIME type
- * `audio/x-speex-with-header-byte`.
- * It is a variant of the RTP Speex encoding defined in
- * [RFC 5574](https://tools.ietf.org/html/rfc5574).
- * The stream is a sequence of blocks, one block per RTP packet. Each block
- * starts with a byte containing the length of the block, in bytes, followed
- * by one or more frames of Speex data, padded to an integral number of
- * bytes (octets) as specified in RFC 5574. In other words, each RTP header
- * is replaced with a single byte containing the block length. Only Speex
- * wideband is supported. `sample_rate_hertz` must be 16000.
- */
+
+ /// *Required* Encoding of audio data sent in all `RecognitionAudio`
+ /// messages.
+ /// Possible string values are:
+ /// - "ENCODING_UNSPECIFIED" : Not specified. Will return result
+ /// google.rpc.Code.INVALID_ARGUMENT.
+ /// - "LINEAR16" : Uncompressed 16-bit signed little-endian samples (Linear
+ /// PCM).
+ /// - "FLAC" : [`FLAC`](https://xiph.org/flac/documentation.html) (Free
+ /// Lossless Audio
+ /// Codec) is the recommended encoding because it is
+ /// lossless--therefore recognition is not compromised--and
+ /// requires only about half the bandwidth of `LINEAR16`. `FLAC` stream
+ /// encoding supports 16-bit and 24-bit samples, however, not all fields in
+ /// `STREAMINFO` are supported.
+ /// - "MULAW" : 8-bit samples that compand 14-bit audio samples using G.711
+ /// PCMU/mu-law.
+ /// - "AMR" : Adaptive Multi-Rate Narrowband codec. `sample_rate_hertz` must
+ /// be 8000.
+ /// - "AMR_WB" : Adaptive Multi-Rate Wideband codec. `sample_rate_hertz` must
+ /// be 16000.
+ /// - "OGG_OPUS" : Opus encoded audio frames in Ogg container
+ /// ([OggOpus](https://wiki.xiph.org/OggOpus)).
+ /// `sample_rate_hertz` must be 16000.
+ /// - "SPEEX_WITH_HEADER_BYTE" : Although the use of lossy encodings is not
+ /// recommended, if a very low
+ /// bitrate encoding is required, `OGG_OPUS` is highly preferred over
+ /// Speex encoding. The [Speex](https://speex.org/) encoding supported by
+ /// Cloud Speech API has a header byte in each block, as in MIME type
+ /// `audio/x-speex-with-header-byte`.
+ /// It is a variant of the RTP Speex encoding defined in
+ /// [RFC 5574](https://tools.ietf.org/html/rfc5574).
+ /// The stream is a sequence of blocks, one block per RTP packet. Each block
+ /// starts with a byte containing the length of the block, in bytes, followed
+ /// by one or more frames of Speex data, padded to an integral number of
+ /// bytes (octets) as specified in RFC 5574. In other words, each RTP header
+ /// is replaced with a single byte containing the block length. Only Speex
+ /// wideband is supported. `sample_rate_hertz` must be 16000.
core.String encoding;
- /**
- * *Required* The language of the supplied audio as a
- * [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
- * Example: "en-US".
- * See [Language Support](https://cloud.google.com/speech/docs/languages)
- * for a list of the currently supported language codes.
- */
+
+ /// *Required* The language of the supplied audio as a
+ /// [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
+ /// Example: "en-US".
+ /// See [Language Support](https://cloud.google.com/speech/docs/languages)
+ /// for a list of the currently supported language codes.
core.String languageCode;
- /**
- * *Optional* Maximum number of recognition hypotheses to be returned.
- * Specifically, the maximum number of `SpeechRecognitionAlternative` messages
- * within each `SpeechRecognitionResult`.
- * The server may return fewer than `max_alternatives`.
- * Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
- * one. If omitted, will return a maximum of one.
- */
+
+ /// *Optional* Maximum number of recognition hypotheses to be returned.
+ /// Specifically, the maximum number of `SpeechRecognitionAlternative`
+ /// messages
+ /// within each `SpeechRecognitionResult`.
+ /// The server may return fewer than `max_alternatives`.
+ /// Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
+ /// one. If omitted, will return a maximum of one.
core.int maxAlternatives;
- /**
- * *Optional* If set to `true`, the server will attempt to filter out
- * profanities, replacing all but the initial character in each filtered word
- * with asterisks, e.g. "f***". If set to `false` or omitted, profanities
- * won't be filtered out.
- */
+
+ /// *Optional* If set to `true`, the server will attempt to filter out
+ /// profanities, replacing all but the initial character in each filtered
+ /// word
+ /// with asterisks, e.g. "f***". If set to `false` or omitted, profanities
+ /// won't be filtered out.
core.bool profanityFilter;
- /**
- * *Required* Sample rate in Hertz of the audio data sent in all
- * `RecognitionAudio` messages. Valid values are: 8000-48000.
- * 16000 is optimal. For best results, set the sampling rate of the audio
- * source to 16000 Hz. If that's not possible, use the native sample rate of
- * the audio source (instead of re-sampling).
- */
+
+ /// *Required* Sample rate in Hertz of the audio data sent in all
+ /// `RecognitionAudio` messages. Valid values are: 8000-48000.
+ /// 16000 is optimal. For best results, set the sampling rate of the audio
+ /// source to 16000 Hz. If that's not possible, use the native sample rate of
+ /// the audio source (instead of re-sampling).
core.int sampleRateHertz;
- /**
- * *Optional* A means to provide context to assist the speech recognition.
- */
+
+ /// *Optional* A means to provide context to assist the speech recognition.
core.List<SpeechContext> speechContexts;
RecognitionConfig();
@@ -685,12 +659,15 @@ class RecognitionConfig {
sampleRateHertz = _json["sampleRateHertz"];
}
if (_json.containsKey("speechContexts")) {
- speechContexts = _json["speechContexts"].map((value) => new SpeechContext.fromJson(value)).toList();
+ speechContexts = _json["speechContexts"]
+ .map((value) => new SpeechContext.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (enableWordTimeOffsets != null) {
_json["enableWordTimeOffsets"] = enableWordTimeOffsets;
}
@@ -710,20 +687,20 @@ class RecognitionConfig {
_json["sampleRateHertz"] = sampleRateHertz;
}
if (speechContexts != null) {
- _json["speechContexts"] = speechContexts.map((value) => (value).toJson()).toList();
+ _json["speechContexts"] =
+ speechContexts.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/** The top-level message sent by the client for the `Recognize` method. */
+/// The top-level message sent by the client for the `Recognize` method.
class RecognizeRequest {
- /** *Required* The audio data to be recognized. */
+ /// *Required* The audio data to be recognized.
RecognitionAudio audio;
- /**
- * *Required* Provides information to the recognizer that specifies how to
- * process the request.
- */
+
+ /// *Required* Provides information to the recognizer that specifies how to
+ /// process the request.
RecognitionConfig config;
RecognizeRequest();
@@ -738,7 +715,8 @@ class RecognizeRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (audio != null) {
_json["audio"] = (audio).toJson();
}
@@ -749,28 +727,27 @@ class RecognizeRequest {
}
}
-/**
- * The only message returned to the client by the `Recognize` method. It
- * contains the result as zero or more sequential `SpeechRecognitionResult`
- * messages.
- */
+/// The only message returned to the client by the `Recognize` method. It
+/// contains the result as zero or more sequential `SpeechRecognitionResult`
+/// messages.
class RecognizeResponse {
- /**
- * *Output-only* Sequential list of transcription results corresponding to
- * sequential portions of audio.
- */
+ /// *Output-only* Sequential list of transcription results corresponding to
+ /// sequential portions of audio.
core.List<SpeechRecognitionResult> results;
RecognizeResponse();
RecognizeResponse.fromJson(core.Map _json) {
if (_json.containsKey("results")) {
- results = _json["results"].map((value) => new SpeechRecognitionResult.fromJson(value)).toList();
+ results = _json["results"]
+ .map((value) => new SpeechRecognitionResult.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (results != null) {
_json["results"] = results.map((value) => (value).toJson()).toList();
}
@@ -778,19 +755,16 @@ class RecognizeResponse {
}
}
-/**
- * Provides "hints" to the speech recognizer to favor specific words and phrases
- * in the results.
- */
+/// Provides "hints" to the speech recognizer to favor specific words and
+/// phrases
+/// in the results.
class SpeechContext {
- /**
- * *Optional* A list of strings containing words and phrases "hints" so that
- * the speech recognition is more likely to recognize them. This can be used
- * to improve the accuracy for specific words and phrases, for example, if
- * specific commands are typically spoken by the user. This can also be used
- * to add additional words to the vocabulary of the recognizer. See
- * [usage limits](https://cloud.google.com/speech/limits#content).
- */
+ /// *Optional* A list of strings containing words and phrases "hints" so that
+ /// the speech recognition is more likely to recognize them. This can be used
+ /// to improve the accuracy for specific words and phrases, for example, if
+ /// specific commands are typically spoken by the user. This can also be used
+ /// to add additional words to the vocabulary of the recognizer. See
+ /// [usage limits](https://cloud.google.com/speech/limits#content).
core.List<core.String> phrases;
SpeechContext();
@@ -802,7 +776,8 @@ class SpeechContext {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (phrases != null) {
_json["phrases"] = phrases;
}
@@ -810,24 +785,24 @@ class SpeechContext {
}
}
-/** Alternative hypotheses (a.k.a. n-best list). */
+/// Alternative hypotheses (a.k.a. n-best list).
class SpeechRecognitionAlternative {
- /**
- * *Output-only* The confidence estimate between 0.0 and 1.0. A higher number
- * indicates an estimated greater likelihood that the recognized words are
- * correct. This field is typically provided only for the top hypothesis, and
- * only for `is_final=true` results. Clients should not rely on the
- * `confidence` field as it is not guaranteed to be accurate or consistent.
- * The default of 0.0 is a sentinel value indicating `confidence` was not set.
- */
+ /// *Output-only* The confidence estimate between 0.0 and 1.0. A higher
+ /// number
+ /// indicates an estimated greater likelihood that the recognized words are
+ /// correct. This field is typically provided only for the top hypothesis,
+ /// and
+ /// only for `is_final=true` results. Clients should not rely on the
+ /// `confidence` field as it is not guaranteed to be accurate or consistent.
+ /// The default of 0.0 is a sentinel value indicating `confidence` was not
+ /// set.
core.double confidence;
- /**
- * *Output-only* Transcript text representing the words that the user spoke.
- */
+
+ /// *Output-only* Transcript text representing the words that the user spoke.
core.String transcript;
- /**
- * *Output-only* A list of word-specific information for each recognized word.
- */
+
+ /// *Output-only* A list of word-specific information for each recognized
+ /// word.
core.List<WordInfo> words;
SpeechRecognitionAlternative();
@@ -840,12 +815,14 @@ class SpeechRecognitionAlternative {
transcript = _json["transcript"];
}
if (_json.containsKey("words")) {
- words = _json["words"].map((value) => new WordInfo.fromJson(value)).toList();
+ words =
+ _json["words"].map((value) => new WordInfo.fromJson(value)).toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (confidence != null) {
_json["confidence"] = confidence;
}
@@ -859,105 +836,110 @@ class SpeechRecognitionAlternative {
}
}
-/** A speech recognition result corresponding to a portion of the audio. */
+/// A speech recognition result corresponding to a portion of the audio.
class SpeechRecognitionResult {
- /**
- * *Output-only* May contain one or more recognition hypotheses (up to the
- * maximum specified in `max_alternatives`).
- * These alternatives are ordered in terms of accuracy, with the top (first)
- * alternative being the most probable, as ranked by the recognizer.
- */
+ /// *Output-only* May contain one or more recognition hypotheses (up to the
+ /// maximum specified in `max_alternatives`).
+ /// These alternatives are ordered in terms of accuracy, with the top (first)
+ /// alternative being the most probable, as ranked by the recognizer.
core.List<SpeechRecognitionAlternative> alternatives;
SpeechRecognitionResult();
SpeechRecognitionResult.fromJson(core.Map _json) {
if (_json.containsKey("alternatives")) {
- alternatives = _json["alternatives"].map((value) => new SpeechRecognitionAlternative.fromJson(value)).toList();
+ alternatives = _json["alternatives"]
+ .map((value) => new SpeechRecognitionAlternative.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (alternatives != null) {
- _json["alternatives"] = alternatives.map((value) => (value).toJson()).toList();
+ _json["alternatives"] =
+ alternatives.map((value) => (value).toJson()).toList();
}
return _json;
}
}
-/**
- * The `Status` type defines a logical error model that is suitable for
- * different
- * programming environments, including REST APIs and RPC APIs. It is used by
- * [gRPC](https://github.com/grpc). The error model is designed to be:
- *
- * - Simple to use and understand for most users
- * - Flexible enough to meet unexpected needs
- *
- * # Overview
- *
- * The `Status` message contains three pieces of data: error code, error
- * message,
- * and error details. The error code should be an enum value of
- * google.rpc.Code, but it may accept additional error codes if needed. The
- * error message should be a developer-facing English message that helps
- * developers *understand* and *resolve* the error. If a localized user-facing
- * error message is needed, put the localized message in the error details or
- * localize it in the client. The optional error details may contain arbitrary
- * information about the error. There is a predefined set of error detail types
- * in the package `google.rpc` that can be used for common error conditions.
- *
- * # Language mapping
- *
- * The `Status` message is the logical representation of the error model, but it
- * is not necessarily the actual wire format. When the `Status` message is
- * exposed in different client libraries and different wire protocols, it can be
- * mapped differently. For example, it will likely be mapped to some exceptions
- * in Java, but more likely mapped to some error codes in C.
- *
- * # Other uses
- *
- * The error model and the `Status` message can be used in a variety of
- * environments, either with or without APIs, to provide a
- * consistent developer experience across different environments.
- *
- * Example uses of this error model include:
- *
- * - Partial errors. If a service needs to return partial errors to the client,
- * it may embed the `Status` in the normal response to indicate the partial
- * errors.
- *
- * - Workflow errors. A typical workflow has multiple steps. Each step may
- * have a `Status` message for error reporting.
- *
- * - Batch operations. If a client uses batch request and batch response, the
- * `Status` message should be used directly inside batch response, one for
- * each error sub-response.
- *
- * - Asynchronous operations. If an API call embeds asynchronous operation
- * results in its response, the status of those operations should be
- * represented directly using the `Status` message.
- *
- * - Logging. If some API errors are stored in logs, the message `Status` could
- * be used directly after any stripping needed for security/privacy reasons.
- */
+/// The `Status` type defines a logical error model that is suitable for
+/// different
+/// programming environments, including REST APIs and RPC APIs. It is used by
+/// [gRPC](https://github.com/grpc). The error model is designed to be:
+///
+/// - Simple to use and understand for most users
+/// - Flexible enough to meet unexpected needs
+///
+/// # Overview
+///
+/// The `Status` message contains three pieces of data: error code, error
+/// message,
+/// and error details. The error code should be an enum value of
+/// google.rpc.Code, but it may accept additional error codes if needed. The
+/// error message should be a developer-facing English message that helps
+/// developers *understand* and *resolve* the error. If a localized user-facing
+/// error message is needed, put the localized message in the error details or
+/// localize it in the client. The optional error details may contain arbitrary
+/// information about the error. There is a predefined set of error detail
+/// types
+/// in the package `google.rpc` that can be used for common error conditions.
+///
+/// # Language mapping
+///
+/// The `Status` message is the logical representation of the error model, but
+/// it
+/// is not necessarily the actual wire format. When the `Status` message is
+/// exposed in different client libraries and different wire protocols, it can
+/// be
+/// mapped differently. For example, it will likely be mapped to some
+/// exceptions
+/// in Java, but more likely mapped to some error codes in C.
+///
+/// # Other uses
+///
+/// The error model and the `Status` message can be used in a variety of
+/// environments, either with or without APIs, to provide a
+/// consistent developer experience across different environments.
+///
+/// Example uses of this error model include:
+///
+/// - Partial errors. If a service needs to return partial errors to the
+/// client,
+/// it may embed the `Status` in the normal response to indicate the partial
+/// errors.
+///
+/// - Workflow errors. A typical workflow has multiple steps. Each step may
+/// have a `Status` message for error reporting.
+///
+/// - Batch operations. If a client uses batch request and batch response, the
+/// `Status` message should be used directly inside batch response, one for
+/// each error sub-response.
+///
+/// - Asynchronous operations. If an API call embeds asynchronous operation
+/// results in its response, the status of those operations should be
+/// represented directly using the `Status` message.
+///
+/// - Logging. If some API errors are stored in logs, the message `Status`
+/// could
+/// be used directly after any stripping needed for security/privacy reasons.
class Status {
- /** The status code, which should be an enum value of google.rpc.Code. */
+ /// The status code, which should be an enum value of google.rpc.Code.
core.int code;
- /**
- * A list of messages that carry the error details. There is a common set of
- * message types for APIs to use.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A list of messages that carry the error details. There is a common set
+ /// of
+ /// message types for APIs to use.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.List<core.Map<core.String, core.Object>> details;
- /**
- * A developer-facing error message, which should be in English. Any
- * user-facing error message should be localized and sent in the
- * google.rpc.Status.details field, or localized by the client.
- */
+
+ /// A developer-facing error message, which should be in English. Any
+ /// user-facing error message should be localized and sent in the
+ /// google.rpc.Status.details field, or localized by the client.
core.String message;
Status();
@@ -975,7 +957,8 @@ class Status {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (code != null) {
_json["code"] = code;
}
@@ -989,31 +972,27 @@ class Status {
}
}
-/**
- * Word-specific information for recognized words. Word information is only
- * included in the response when certain request parameters are set, such
- * as `enable_word_time_offsets`.
- */
+/// Word-specific information for recognized words. Word information is only
+/// included in the response when certain request parameters are set, such
+/// as `enable_word_time_offsets`.
class WordInfo {
- /**
- * *Output-only* Time offset relative to the beginning of the audio,
- * and corresponding to the end of the spoken word.
- * This field is only set if `enable_word_time_offsets=true` and only
- * in the top hypothesis.
- * This is an experimental feature and the accuracy of the time offset can
- * vary.
- */
+ /// *Output-only* Time offset relative to the beginning of the audio,
+ /// and corresponding to the end of the spoken word.
+ /// This field is only set if `enable_word_time_offsets=true` and only
+ /// in the top hypothesis.
+ /// This is an experimental feature and the accuracy of the time offset can
+ /// vary.
core.String endTime;
- /**
- * *Output-only* Time offset relative to the beginning of the audio,
- * and corresponding to the start of the spoken word.
- * This field is only set if `enable_word_time_offsets=true` and only
- * in the top hypothesis.
- * This is an experimental feature and the accuracy of the time offset can
- * vary.
- */
+
+ /// *Output-only* Time offset relative to the beginning of the audio,
+ /// and corresponding to the start of the spoken word.
+ /// This field is only set if `enable_word_time_offsets=true` and only
+ /// in the top hypothesis.
+ /// This is an experimental feature and the accuracy of the time offset can
+ /// vary.
core.String startTime;
- /** *Output-only* The word corresponding to this set of information. */
+
+ /// *Output-only* The word corresponding to this set of information.
core.String word;
WordInfo();
@@ -1031,7 +1010,8 @@ class WordInfo {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (endTime != null) {
_json["endTime"] = endTime;
}
« no previous file with comments | « generated/googleapis/lib/spanner/v1.dart ('k') | generated/googleapis/lib/storage/v1.dart » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698