Chromium Code Reviews
chromiumcodereview-hr@appspot.gserviceaccount.com (chromiumcodereview-hr) | Please choose your nickname with Settings | Help | Chromium Project | Gerrit Changes | Sign out
(115)

Unified Diff: generated/googleapis/lib/language/v1.dart

Issue 3006323002: Api-Roll 54: 2017-09-11 (Closed)
Patch Set: use 2.0.0-dev.infinity sdk constraint in pubspecs Created 3 years, 3 months ago
Use n/p to move between diff chunks; N/P to move between comments. Draft comments are only viewable by you.
Jump to:
View side-by-side diff with in-line comments
Download patch
« no previous file with comments | « generated/googleapis/lib/kgsearch/v1.dart ('k') | generated/googleapis/lib/licensing/v1.dart » ('j') | no next file with comments »
Expand Comments ('e') | Collapse Comments ('c') | Show Comments Hide Comments ('s')
Index: generated/googleapis/lib/language/v1.dart
diff --git a/generated/googleapis/lib/language/v1.dart b/generated/googleapis/lib/language/v1.dart
index 6491348a7d31779ebbcd5f788160e9b07f410f25..642544c04a863eb13cfac9f24bfc53253d169364 100644
--- a/generated/googleapis/lib/language/v1.dart
+++ b/generated/googleapis/lib/language/v1.dart
@@ -9,59 +9,57 @@ import 'dart:convert' as convert;
import 'package:_discoveryapis_commons/_discoveryapis_commons.dart' as commons;
import 'package:http/http.dart' as http;
-export 'package:_discoveryapis_commons/_discoveryapis_commons.dart' show
- ApiRequestError, DetailedApiRequestError;
+export 'package:_discoveryapis_commons/_discoveryapis_commons.dart'
+ show ApiRequestError, DetailedApiRequestError;
const core.String USER_AGENT = 'dart-api-client language/v1';
-/**
- * Provides natural language understanding technologies to developers. Examples
- * include sentiment analysis, entity recognition, entity sentiment analysis,
- * and text annotations.
- */
+/// Provides natural language understanding technologies to developers.
+/// Examples include sentiment analysis, entity recognition, entity sentiment
+/// analysis, and text annotations.
class LanguageApi {
- /**
- * Apply machine learning models to reveal the structure and meaning of text
- */
- static const CloudLanguageScope = "https://www.googleapis.com/auth/cloud-language";
-
- /** View and manage your data across Google Cloud Platform services */
- static const CloudPlatformScope = "https://www.googleapis.com/auth/cloud-platform";
+ /// Apply machine learning models to reveal the structure and meaning of text
+ static const CloudLanguageScope =
+ "https://www.googleapis.com/auth/cloud-language";
+ /// View and manage your data across Google Cloud Platform services
+ static const CloudPlatformScope =
+ "https://www.googleapis.com/auth/cloud-platform";
final commons.ApiRequester _requester;
DocumentsResourceApi get documents => new DocumentsResourceApi(_requester);
- LanguageApi(http.Client client, {core.String rootUrl: "https://language.googleapis.com/", core.String servicePath: ""}) :
- _requester = new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
+ LanguageApi(http.Client client,
+ {core.String rootUrl: "https://language.googleapis.com/",
+ core.String servicePath: ""})
+ : _requester =
+ new commons.ApiRequester(client, rootUrl, servicePath, USER_AGENT);
}
-
class DocumentsResourceApi {
final commons.ApiRequester _requester;
- DocumentsResourceApi(commons.ApiRequester client) :
- _requester = client;
-
- /**
- * Finds named entities (currently proper names and common nouns) in the text
- * along with entity types, salience, mentions for each entity, and
- * other properties.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [AnalyzeEntitiesResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<AnalyzeEntitiesResponse> analyzeEntities(AnalyzeEntitiesRequest request) {
+ DocumentsResourceApi(commons.ApiRequester client) : _requester = client;
+
+ /// Finds named entities (currently proper names and common nouns) in the
+ /// text
+ /// along with entity types, salience, mentions for each entity, and
+ /// other properties.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [AnalyzeEntitiesResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<AnalyzeEntitiesResponse> analyzeEntities(
+ AnalyzeEntitiesRequest request) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -75,32 +73,30 @@ class DocumentsResourceApi {
_url = 'v1/documents:analyzeEntities';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new AnalyzeEntitiesResponse.fromJson(data));
}
- /**
- * Analyzes the sentiment of the provided text.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [AnalyzeSentimentResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<AnalyzeSentimentResponse> analyzeSentiment(AnalyzeSentimentRequest request) {
+ /// Analyzes the sentiment of the provided text.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [AnalyzeSentimentResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<AnalyzeSentimentResponse> analyzeSentiment(
+ AnalyzeSentimentRequest request) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -114,34 +110,33 @@ class DocumentsResourceApi {
_url = 'v1/documents:analyzeSentiment';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
- return _response.then((data) => new AnalyzeSentimentResponse.fromJson(data));
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
+ return _response
+ .then((data) => new AnalyzeSentimentResponse.fromJson(data));
}
- /**
- * Analyzes the syntax of the text and provides sentence boundaries and
- * tokenization along with part of speech tags, dependency trees, and other
- * properties.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [AnalyzeSyntaxResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
- async.Future<AnalyzeSyntaxResponse> analyzeSyntax(AnalyzeSyntaxRequest request) {
+ /// Analyzes the syntax of the text and provides sentence boundaries and
+ /// tokenization along with part of speech tags, dependency trees, and other
+ /// properties.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [AnalyzeSyntaxResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
+ async.Future<AnalyzeSyntaxResponse> analyzeSyntax(
+ AnalyzeSyntaxRequest request) {
var _url = null;
var _queryParams = new core.Map();
var _uploadMedia = null;
@@ -155,32 +150,30 @@ class DocumentsResourceApi {
_url = 'v1/documents:analyzeSyntax';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new AnalyzeSyntaxResponse.fromJson(data));
}
- /**
- * A convenience method that provides all the features that analyzeSentiment,
- * analyzeEntities, and analyzeSyntax provide in one call.
- *
- * [request] - The metadata request object.
- *
- * Request parameters:
- *
- * Completes with a [AnnotateTextResponse].
- *
- * Completes with a [commons.ApiRequestError] if the API endpoint returned an
- * error.
- *
- * If the used [http.Client] completes with an error when making a REST call,
- * this method will complete with the same error.
- */
+ /// A convenience method that provides all the features that
+ /// analyzeSentiment,
+ /// analyzeEntities, and analyzeSyntax provide in one call.
+ ///
+ /// [request] - The metadata request object.
+ ///
+ /// Request parameters:
+ ///
+ /// Completes with a [AnnotateTextResponse].
+ ///
+ /// Completes with a [commons.ApiRequestError] if the API endpoint returned
+ /// an error.
+ ///
+ /// If the used [http.Client] completes with an error when making a REST
+ /// call, this method will complete with the same error.
async.Future<AnnotateTextResponse> annotateText(AnnotateTextRequest request) {
var _url = null;
var _queryParams = new core.Map();
@@ -195,43 +188,38 @@ class DocumentsResourceApi {
_url = 'v1/documents:annotateText';
- var _response = _requester.request(_url,
- "POST",
- body: _body,
- queryParams: _queryParams,
- uploadOptions: _uploadOptions,
- uploadMedia: _uploadMedia,
- downloadOptions: _downloadOptions);
+ var _response = _requester.request(_url, "POST",
+ body: _body,
+ queryParams: _queryParams,
+ uploadOptions: _uploadOptions,
+ uploadMedia: _uploadMedia,
+ downloadOptions: _downloadOptions);
return _response.then((data) => new AnnotateTextResponse.fromJson(data));
}
-
}
-
-
-/** The entity analysis request message. */
+/// The entity analysis request message.
class AnalyzeEntitiesRequest {
- /** Input document. */
+ /// Input document.
Document document;
- /**
- * The encoding type used by the API to calculate offsets.
- * Possible string values are:
- * - "NONE" : If `EncodingType` is not specified, encoding-dependent
- * information (such as
- * `begin_offset`) will be set at `-1`.
- * - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-8 encoding of the input. C++ and Go are examples of languages
- * that use this encoding natively.
- * - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-16 encoding of the input. Java and Javascript are examples of
- * languages that use this encoding natively.
- * - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-32 encoding of the input. Python is an example of a language
- * that uses this encoding natively.
- */
+
+ /// The encoding type used by the API to calculate offsets.
+ /// Possible string values are:
+ /// - "NONE" : If `EncodingType` is not specified, encoding-dependent
+ /// information (such as
+ /// `begin_offset`) will be set at `-1`.
+ /// - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-8 encoding of the input. C++ and Go are examples of languages
+ /// that use this encoding natively.
+ /// - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-16 encoding of the input. Java and Javascript are examples of
+ /// languages that use this encoding natively.
+ /// - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-32 encoding of the input. Python is an example of a language
+ /// that uses this encoding natively.
core.String encodingType;
AnalyzeEntitiesRequest();
@@ -246,7 +234,8 @@ class AnalyzeEntitiesRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (document != null) {
_json["document"] = (document).toJson();
}
@@ -257,22 +246,23 @@ class AnalyzeEntitiesRequest {
}
}
-/** The entity analysis response message. */
+/// The entity analysis response message.
class AnalyzeEntitiesResponse {
- /** The recognized entities in the input document. */
+ /// The recognized entities in the input document.
core.List<Entity> entities;
- /**
- * The language of the text, which will be the same as the language specified
- * in the request or, if not specified, the automatically-detected language.
- * See Document.language field for more details.
- */
+
+ /// The language of the text, which will be the same as the language
+ /// specified
+ /// in the request or, if not specified, the automatically-detected language.
+ /// See Document.language field for more details.
core.String language;
AnalyzeEntitiesResponse();
AnalyzeEntitiesResponse.fromJson(core.Map _json) {
if (_json.containsKey("entities")) {
- entities = _json["entities"].map((value) => new Entity.fromJson(value)).toList();
+ entities =
+ _json["entities"].map((value) => new Entity.fromJson(value)).toList();
}
if (_json.containsKey("language")) {
language = _json["language"];
@@ -280,7 +270,8 @@ class AnalyzeEntitiesResponse {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (entities != null) {
_json["entities"] = entities.map((value) => (value).toJson()).toList();
}
@@ -291,29 +282,28 @@ class AnalyzeEntitiesResponse {
}
}
-/** The sentiment analysis request message. */
+/// The sentiment analysis request message.
class AnalyzeSentimentRequest {
- /** Input document. */
+ /// Input document.
Document document;
- /**
- * The encoding type used by the API to calculate sentence offsets.
- * Possible string values are:
- * - "NONE" : If `EncodingType` is not specified, encoding-dependent
- * information (such as
- * `begin_offset`) will be set at `-1`.
- * - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-8 encoding of the input. C++ and Go are examples of languages
- * that use this encoding natively.
- * - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-16 encoding of the input. Java and Javascript are examples of
- * languages that use this encoding natively.
- * - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-32 encoding of the input. Python is an example of a language
- * that uses this encoding natively.
- */
+
+ /// The encoding type used by the API to calculate sentence offsets.
+ /// Possible string values are:
+ /// - "NONE" : If `EncodingType` is not specified, encoding-dependent
+ /// information (such as
+ /// `begin_offset`) will be set at `-1`.
+ /// - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-8 encoding of the input. C++ and Go are examples of languages
+ /// that use this encoding natively.
+ /// - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-16 encoding of the input. Java and Javascript are examples of
+ /// languages that use this encoding natively.
+ /// - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-32 encoding of the input. Python is an example of a language
+ /// that uses this encoding natively.
core.String encodingType;
AnalyzeSentimentRequest();
@@ -328,7 +318,8 @@ class AnalyzeSentimentRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (document != null) {
_json["document"] = (document).toJson();
}
@@ -339,17 +330,18 @@ class AnalyzeSentimentRequest {
}
}
-/** The sentiment analysis response message. */
+/// The sentiment analysis response message.
class AnalyzeSentimentResponse {
- /** The overall sentiment of the input document. */
+ /// The overall sentiment of the input document.
Sentiment documentSentiment;
- /**
- * The language of the text, which will be the same as the language specified
- * in the request or, if not specified, the automatically-detected language.
- * See Document.language field for more details.
- */
+
+ /// The language of the text, which will be the same as the language
+ /// specified
+ /// in the request or, if not specified, the automatically-detected language.
+ /// See Document.language field for more details.
core.String language;
- /** The sentiment for all the sentences in the document. */
+
+ /// The sentiment for all the sentences in the document.
core.List<Sentence> sentences;
AnalyzeSentimentResponse();
@@ -362,12 +354,15 @@ class AnalyzeSentimentResponse {
language = _json["language"];
}
if (_json.containsKey("sentences")) {
- sentences = _json["sentences"].map((value) => new Sentence.fromJson(value)).toList();
+ sentences = _json["sentences"]
+ .map((value) => new Sentence.fromJson(value))
+ .toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (documentSentiment != null) {
_json["documentSentiment"] = (documentSentiment).toJson();
}
@@ -381,29 +376,28 @@ class AnalyzeSentimentResponse {
}
}
-/** The syntax analysis request message. */
+/// The syntax analysis request message.
class AnalyzeSyntaxRequest {
- /** Input document. */
+ /// Input document.
Document document;
- /**
- * The encoding type used by the API to calculate offsets.
- * Possible string values are:
- * - "NONE" : If `EncodingType` is not specified, encoding-dependent
- * information (such as
- * `begin_offset`) will be set at `-1`.
- * - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-8 encoding of the input. C++ and Go are examples of languages
- * that use this encoding natively.
- * - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-16 encoding of the input. Java and Javascript are examples of
- * languages that use this encoding natively.
- * - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-32 encoding of the input. Python is an example of a language
- * that uses this encoding natively.
- */
+
+ /// The encoding type used by the API to calculate offsets.
+ /// Possible string values are:
+ /// - "NONE" : If `EncodingType` is not specified, encoding-dependent
+ /// information (such as
+ /// `begin_offset`) will be set at `-1`.
+ /// - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-8 encoding of the input. C++ and Go are examples of languages
+ /// that use this encoding natively.
+ /// - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-16 encoding of the input. Java and Javascript are examples of
+ /// languages that use this encoding natively.
+ /// - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-32 encoding of the input. Python is an example of a language
+ /// that uses this encoding natively.
core.String encodingType;
AnalyzeSyntaxRequest();
@@ -418,7 +412,8 @@ class AnalyzeSyntaxRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (document != null) {
_json["document"] = (document).toJson();
}
@@ -429,17 +424,18 @@ class AnalyzeSyntaxRequest {
}
}
-/** The syntax analysis response message. */
+/// The syntax analysis response message.
class AnalyzeSyntaxResponse {
- /**
- * The language of the text, which will be the same as the language specified
- * in the request or, if not specified, the automatically-detected language.
- * See Document.language field for more details.
- */
+ /// The language of the text, which will be the same as the language
+ /// specified
+ /// in the request or, if not specified, the automatically-detected language.
+ /// See Document.language field for more details.
core.String language;
- /** Sentences in the input document. */
+
+ /// Sentences in the input document.
core.List<Sentence> sentences;
- /** Tokens, along with their syntactic information, in the input document. */
+
+ /// Tokens, along with their syntactic information, in the input document.
core.List<Token> tokens;
AnalyzeSyntaxResponse();
@@ -449,15 +445,19 @@ class AnalyzeSyntaxResponse {
language = _json["language"];
}
if (_json.containsKey("sentences")) {
- sentences = _json["sentences"].map((value) => new Sentence.fromJson(value)).toList();
+ sentences = _json["sentences"]
+ .map((value) => new Sentence.fromJson(value))
+ .toList();
}
if (_json.containsKey("tokens")) {
- tokens = _json["tokens"].map((value) => new Token.fromJson(value)).toList();
+ tokens =
+ _json["tokens"].map((value) => new Token.fromJson(value)).toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (language != null) {
_json["language"] = language;
}
@@ -471,34 +471,32 @@ class AnalyzeSyntaxResponse {
}
}
-/**
- * The request message for the text annotation API, which can perform multiple
- * analysis types (sentiment, entities, and syntax) in one call.
- */
+/// The request message for the text annotation API, which can perform multiple
+/// analysis types (sentiment, entities, and syntax) in one call.
class AnnotateTextRequest {
- /** Input document. */
+ /// Input document.
Document document;
- /**
- * The encoding type used by the API to calculate offsets.
- * Possible string values are:
- * - "NONE" : If `EncodingType` is not specified, encoding-dependent
- * information (such as
- * `begin_offset`) will be set at `-1`.
- * - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-8 encoding of the input. C++ and Go are examples of languages
- * that use this encoding natively.
- * - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-16 encoding of the input. Java and Javascript are examples of
- * languages that use this encoding natively.
- * - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
- * calculated based
- * on the UTF-32 encoding of the input. Python is an example of a language
- * that uses this encoding natively.
- */
+
+ /// The encoding type used by the API to calculate offsets.
+ /// Possible string values are:
+ /// - "NONE" : If `EncodingType` is not specified, encoding-dependent
+ /// information (such as
+ /// `begin_offset`) will be set at `-1`.
+ /// - "UTF8" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-8 encoding of the input. C++ and Go are examples of languages
+ /// that use this encoding natively.
+ /// - "UTF16" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-16 encoding of the input. Java and Javascript are examples of
+ /// languages that use this encoding natively.
+ /// - "UTF32" : Encoding-dependent information (such as `begin_offset`) is
+ /// calculated based
+ /// on the UTF-32 encoding of the input. Python is an example of a language
+ /// that uses this encoding natively.
core.String encodingType;
- /** The enabled features. */
+
+ /// The enabled features.
Features features;
AnnotateTextRequest();
@@ -516,7 +514,8 @@ class AnnotateTextRequest {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (document != null) {
_json["document"] = (document).toJson();
}
@@ -530,35 +529,30 @@ class AnnotateTextRequest {
}
}
-/** The text annotations response message. */
+/// The text annotations response message.
class AnnotateTextResponse {
- /**
- * The overall sentiment for the document. Populated if the user enables
- * AnnotateTextRequest.Features.extract_document_sentiment.
- */
+ /// The overall sentiment for the document. Populated if the user enables
+ /// AnnotateTextRequest.Features.extract_document_sentiment.
Sentiment documentSentiment;
- /**
- * Entities, along with their semantic information, in the input document.
- * Populated if the user enables
- * AnnotateTextRequest.Features.extract_entities.
- */
+
+ /// Entities, along with their semantic information, in the input document.
+ /// Populated if the user enables
+ /// AnnotateTextRequest.Features.extract_entities.
core.List<Entity> entities;
- /**
- * The language of the text, which will be the same as the language specified
- * in the request or, if not specified, the automatically-detected language.
- * See Document.language field for more details.
- */
+
+ /// The language of the text, which will be the same as the language
+ /// specified
+ /// in the request or, if not specified, the automatically-detected language.
+ /// See Document.language field for more details.
core.String language;
- /**
- * Sentences in the input document. Populated if the user enables
- * AnnotateTextRequest.Features.extract_syntax.
- */
+
+ /// Sentences in the input document. Populated if the user enables
+ /// AnnotateTextRequest.Features.extract_syntax.
core.List<Sentence> sentences;
- /**
- * Tokens, along with their syntactic information, in the input document.
- * Populated if the user enables
- * AnnotateTextRequest.Features.extract_syntax.
- */
+
+ /// Tokens, along with their syntactic information, in the input document.
+ /// Populated if the user enables
+ /// AnnotateTextRequest.Features.extract_syntax.
core.List<Token> tokens;
AnnotateTextResponse();
@@ -568,21 +562,26 @@ class AnnotateTextResponse {
documentSentiment = new Sentiment.fromJson(_json["documentSentiment"]);
}
if (_json.containsKey("entities")) {
- entities = _json["entities"].map((value) => new Entity.fromJson(value)).toList();
+ entities =
+ _json["entities"].map((value) => new Entity.fromJson(value)).toList();
}
if (_json.containsKey("language")) {
language = _json["language"];
}
if (_json.containsKey("sentences")) {
- sentences = _json["sentences"].map((value) => new Sentence.fromJson(value)).toList();
+ sentences = _json["sentences"]
+ .map((value) => new Sentence.fromJson(value))
+ .toList();
}
if (_json.containsKey("tokens")) {
- tokens = _json["tokens"].map((value) => new Token.fromJson(value)).toList();
+ tokens =
+ _json["tokens"].map((value) => new Token.fromJson(value)).toList();
}
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (documentSentiment != null) {
_json["documentSentiment"] = (documentSentiment).toJson();
}
@@ -602,102 +601,97 @@ class AnnotateTextResponse {
}
}
-/**
- * Represents dependency parse tree information for a token. (For more
- * information on dependency labels, see
- * http://www.aclweb.org/anthology/P13-2017
- */
+/// Represents dependency parse tree information for a token. (For more
+/// information on dependency labels, see
+/// http://www.aclweb.org/anthology/P13-2017
class DependencyEdge {
- /**
- * Represents the head of this token in the dependency tree.
- * This is the index of the token which has an arc going to this token.
- * The index is the position of the token in the array of tokens returned
- * by the API method. If this token is a root token, then the
- * `head_token_index` is its own index.
- */
+ /// Represents the head of this token in the dependency tree.
+ /// This is the index of the token which has an arc going to this token.
+ /// The index is the position of the token in the array of tokens returned
+ /// by the API method. If this token is a root token, then the
+ /// `head_token_index` is its own index.
core.int headTokenIndex;
- /**
- * The parse label for the token.
- * Possible string values are:
- * - "UNKNOWN" : Unknown
- * - "ABBREV" : Abbreviation modifier
- * - "ACOMP" : Adjectival complement
- * - "ADVCL" : Adverbial clause modifier
- * - "ADVMOD" : Adverbial modifier
- * - "AMOD" : Adjectival modifier of an NP
- * - "APPOS" : Appositional modifier of an NP
- * - "ATTR" : Attribute dependent of a copular verb
- * - "AUX" : Auxiliary (non-main) verb
- * - "AUXPASS" : Passive auxiliary
- * - "CC" : Coordinating conjunction
- * - "CCOMP" : Clausal complement of a verb or adjective
- * - "CONJ" : Conjunct
- * - "CSUBJ" : Clausal subject
- * - "CSUBJPASS" : Clausal passive subject
- * - "DEP" : Dependency (unable to determine)
- * - "DET" : Determiner
- * - "DISCOURSE" : Discourse
- * - "DOBJ" : Direct object
- * - "EXPL" : Expletive
- * - "GOESWITH" : Goes with (part of a word in a text not well edited)
- * - "IOBJ" : Indirect object
- * - "MARK" : Marker (word introducing a subordinate clause)
- * - "MWE" : Multi-word expression
- * - "MWV" : Multi-word verbal expression
- * - "NEG" : Negation modifier
- * - "NN" : Noun compound modifier
- * - "NPADVMOD" : Noun phrase used as an adverbial modifier
- * - "NSUBJ" : Nominal subject
- * - "NSUBJPASS" : Passive nominal subject
- * - "NUM" : Numeric modifier of a noun
- * - "NUMBER" : Element of compound number
- * - "P" : Punctuation mark
- * - "PARATAXIS" : Parataxis relation
- * - "PARTMOD" : Participial modifier
- * - "PCOMP" : The complement of a preposition is a clause
- * - "POBJ" : Object of a preposition
- * - "POSS" : Possession modifier
- * - "POSTNEG" : Postverbal negative particle
- * - "PRECOMP" : Predicate complement
- * - "PRECONJ" : Preconjunt
- * - "PREDET" : Predeterminer
- * - "PREF" : Prefix
- * - "PREP" : Prepositional modifier
- * - "PRONL" : The relationship between a verb and verbal morpheme
- * - "PRT" : Particle
- * - "PS" : Associative or possessive marker
- * - "QUANTMOD" : Quantifier phrase modifier
- * - "RCMOD" : Relative clause modifier
- * - "RCMODREL" : Complementizer in relative clause
- * - "RDROP" : Ellipsis without a preceding predicate
- * - "REF" : Referent
- * - "REMNANT" : Remnant
- * - "REPARANDUM" : Reparandum
- * - "ROOT" : Root
- * - "SNUM" : Suffix specifying a unit of number
- * - "SUFF" : Suffix
- * - "TMOD" : Temporal modifier
- * - "TOPIC" : Topic marker
- * - "VMOD" : Clause headed by an infinite form of the verb that modifies a
- * noun
- * - "VOCATIVE" : Vocative
- * - "XCOMP" : Open clausal complement
- * - "SUFFIX" : Name suffix
- * - "TITLE" : Name title
- * - "ADVPHMOD" : Adverbial phrase modifier
- * - "AUXCAUS" : Causative auxiliary
- * - "AUXVV" : Helper auxiliary
- * - "DTMOD" : Rentaishi (Prenominal modifier)
- * - "FOREIGN" : Foreign words
- * - "KW" : Keyword
- * - "LIST" : List for chains of comparable items
- * - "NOMC" : Nominalized clause
- * - "NOMCSUBJ" : Nominalized clausal subject
- * - "NOMCSUBJPASS" : Nominalized clausal passive
- * - "NUMC" : Compound of numeric modifier
- * - "COP" : Copula
- * - "DISLOCATED" : Dislocated relation (for fronted/topicalized elements)
- */
+
+ /// The parse label for the token.
+ /// Possible string values are:
+ /// - "UNKNOWN" : Unknown
+ /// - "ABBREV" : Abbreviation modifier
+ /// - "ACOMP" : Adjectival complement
+ /// - "ADVCL" : Adverbial clause modifier
+ /// - "ADVMOD" : Adverbial modifier
+ /// - "AMOD" : Adjectival modifier of an NP
+ /// - "APPOS" : Appositional modifier of an NP
+ /// - "ATTR" : Attribute dependent of a copular verb
+ /// - "AUX" : Auxiliary (non-main) verb
+ /// - "AUXPASS" : Passive auxiliary
+ /// - "CC" : Coordinating conjunction
+ /// - "CCOMP" : Clausal complement of a verb or adjective
+ /// - "CONJ" : Conjunct
+ /// - "CSUBJ" : Clausal subject
+ /// - "CSUBJPASS" : Clausal passive subject
+ /// - "DEP" : Dependency (unable to determine)
+ /// - "DET" : Determiner
+ /// - "DISCOURSE" : Discourse
+ /// - "DOBJ" : Direct object
+ /// - "EXPL" : Expletive
+ /// - "GOESWITH" : Goes with (part of a word in a text not well edited)
+ /// - "IOBJ" : Indirect object
+ /// - "MARK" : Marker (word introducing a subordinate clause)
+ /// - "MWE" : Multi-word expression
+ /// - "MWV" : Multi-word verbal expression
+ /// - "NEG" : Negation modifier
+ /// - "NN" : Noun compound modifier
+ /// - "NPADVMOD" : Noun phrase used as an adverbial modifier
+ /// - "NSUBJ" : Nominal subject
+ /// - "NSUBJPASS" : Passive nominal subject
+ /// - "NUM" : Numeric modifier of a noun
+ /// - "NUMBER" : Element of compound number
+ /// - "P" : Punctuation mark
+ /// - "PARATAXIS" : Parataxis relation
+ /// - "PARTMOD" : Participial modifier
+ /// - "PCOMP" : The complement of a preposition is a clause
+ /// - "POBJ" : Object of a preposition
+ /// - "POSS" : Possession modifier
+ /// - "POSTNEG" : Postverbal negative particle
+ /// - "PRECOMP" : Predicate complement
+ /// - "PRECONJ" : Preconjunt
+ /// - "PREDET" : Predeterminer
+ /// - "PREF" : Prefix
+ /// - "PREP" : Prepositional modifier
+ /// - "PRONL" : The relationship between a verb and verbal morpheme
+ /// - "PRT" : Particle
+ /// - "PS" : Associative or possessive marker
+ /// - "QUANTMOD" : Quantifier phrase modifier
+ /// - "RCMOD" : Relative clause modifier
+ /// - "RCMODREL" : Complementizer in relative clause
+ /// - "RDROP" : Ellipsis without a preceding predicate
+ /// - "REF" : Referent
+ /// - "REMNANT" : Remnant
+ /// - "REPARANDUM" : Reparandum
+ /// - "ROOT" : Root
+ /// - "SNUM" : Suffix specifying a unit of number
+ /// - "SUFF" : Suffix
+ /// - "TMOD" : Temporal modifier
+ /// - "TOPIC" : Topic marker
+ /// - "VMOD" : Clause headed by an infinite form of the verb that modifies a
+ /// noun
+ /// - "VOCATIVE" : Vocative
+ /// - "XCOMP" : Open clausal complement
+ /// - "SUFFIX" : Name suffix
+ /// - "TITLE" : Name title
+ /// - "ADVPHMOD" : Adverbial phrase modifier
+ /// - "AUXCAUS" : Causative auxiliary
+ /// - "AUXVV" : Helper auxiliary
+ /// - "DTMOD" : Rentaishi (Prenominal modifier)
+ /// - "FOREIGN" : Foreign words
+ /// - "KW" : Keyword
+ /// - "LIST" : List for chains of comparable items
+ /// - "NOMC" : Nominalized clause
+ /// - "NOMCSUBJ" : Nominalized clausal subject
+ /// - "NOMCSUBJPASS" : Nominalized clausal passive
+ /// - "NUMC" : Compound of numeric modifier
+ /// - "COP" : Copula
+ /// - "DISLOCATED" : Dislocated relation (for fronted/topicalized elements)
core.String label;
DependencyEdge();
@@ -712,7 +706,8 @@ class DependencyEdge {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (headTokenIndex != null) {
_json["headTokenIndex"] = headTokenIndex;
}
@@ -723,40 +718,36 @@ class DependencyEdge {
}
}
-/**
- * ################################################################ #
- *
- * Represents the input to API methods.
- */
+/// ################################################################ #
+///
+/// Represents the input to API methods.
class Document {
- /** The content of the input in string format. */
+ /// The content of the input in string format.
core.String content;
- /**
- * The Google Cloud Storage URI where the file content is located.
- * This URI must be of the form: gs://bucket_name/object_name. For more
- * details, see https://cloud.google.com/storage/docs/reference-uris.
- * NOTE: Cloud Storage object versioning is not supported.
- */
+
+ /// The Google Cloud Storage URI where the file content is located.
+ /// This URI must be of the form: gs://bucket_name/object_name. For more
+ /// details, see https://cloud.google.com/storage/docs/reference-uris.
+ /// NOTE: Cloud Storage object versioning is not supported.
core.String gcsContentUri;
- /**
- * The language of the document (if not specified, the language is
- * automatically detected). Both ISO and BCP-47 language codes are
- * accepted.<br>
- * [Language Support](/natural-language/docs/languages)
- * lists currently supported languages for each API method.
- * If the language (either specified by the caller or automatically detected)
- * is not supported by the called API method, an `INVALID_ARGUMENT` error
- * is returned.
- */
+
+ /// The language of the document (if not specified, the language is
+ /// automatically detected). Both ISO and BCP-47 language codes are
+ /// accepted.<br>
+ /// [Language Support](/natural-language/docs/languages)
+ /// lists currently supported languages for each API method.
+ /// If the language (either specified by the caller or automatically
+ /// detected)
+ /// is not supported by the called API method, an `INVALID_ARGUMENT` error
+ /// is returned.
core.String language;
- /**
- * Required. If the type is not set or is `TYPE_UNSPECIFIED`,
- * returns an `INVALID_ARGUMENT` error.
- * Possible string values are:
- * - "TYPE_UNSPECIFIED" : The content type is not specified.
- * - "PLAIN_TEXT" : Plain text
- * - "HTML" : HTML
- */
+
+ /// Required. If the type is not set or is `TYPE_UNSPECIFIED`,
+ /// returns an `INVALID_ARGUMENT` error.
+ /// Possible string values are:
+ /// - "TYPE_UNSPECIFIED" : The content type is not specified.
+ /// - "PLAIN_TEXT" : Plain text
+ /// - "HTML" : HTML
core.String type;
Document();
@@ -777,7 +768,8 @@ class Document {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (content != null) {
_json["content"] = content;
}
@@ -794,54 +786,53 @@ class Document {
}
}
-/**
- * Represents a phrase in the text that is a known entity, such as
- * a person, an organization, or location. The API associates information, such
- * as salience and mentions, with entities.
- */
+/// Represents a phrase in the text that is a known entity, such as
+/// a person, an organization, or location. The API associates information,
+/// such
+/// as salience and mentions, with entities.
class Entity {
- /**
- * The mentions of this entity in the input document. The API currently
- * supports proper noun mentions.
- */
+ /// The mentions of this entity in the input document. The API currently
+ /// supports proper noun mentions.
core.List<EntityMention> mentions;
- /**
- * Metadata associated with the entity.
- *
- * Currently, Wikipedia URLs and Knowledge Graph MIDs are provided, if
- * available. The associated keys are "wikipedia_url" and "mid", respectively.
- */
+
+ /// Metadata associated with the entity.
+ ///
+ /// Currently, Wikipedia URLs and Knowledge Graph MIDs are provided, if
+ /// available. The associated keys are "wikipedia_url" and "mid",
+ /// respectively.
core.Map<core.String, core.String> metadata;
- /** The representative name for the entity. */
+
+ /// The representative name for the entity.
core.String name;
- /**
- * The salience score associated with the entity in the [0, 1.0] range.
- *
- * The salience score for an entity provides information about the
- * importance or centrality of that entity to the entire document text.
- * Scores closer to 0 are less salient, while scores closer to 1.0 are highly
- * salient.
- */
+
+ /// The salience score associated with the entity in the [0, 1.0] range.
+ ///
+ /// The salience score for an entity provides information about the
+ /// importance or centrality of that entity to the entire document text.
+ /// Scores closer to 0 are less salient, while scores closer to 1.0 are
+ /// highly
+ /// salient.
core.double salience;
- /**
- * The entity type.
- * Possible string values are:
- * - "UNKNOWN" : Unknown
- * - "PERSON" : Person
- * - "LOCATION" : Location
- * - "ORGANIZATION" : Organization
- * - "EVENT" : Event
- * - "WORK_OF_ART" : Work of art
- * - "CONSUMER_GOOD" : Consumer goods
- * - "OTHER" : Other types
- */
+
+ /// The entity type.
+ /// Possible string values are:
+ /// - "UNKNOWN" : Unknown
+ /// - "PERSON" : Person
+ /// - "LOCATION" : Location
+ /// - "ORGANIZATION" : Organization
+ /// - "EVENT" : Event
+ /// - "WORK_OF_ART" : Work of art
+ /// - "CONSUMER_GOOD" : Consumer goods
+ /// - "OTHER" : Other types
core.String type;
Entity();
Entity.fromJson(core.Map _json) {
if (_json.containsKey("mentions")) {
- mentions = _json["mentions"].map((value) => new EntityMention.fromJson(value)).toList();
+ mentions = _json["mentions"]
+ .map((value) => new EntityMention.fromJson(value))
+ .toList();
}
if (_json.containsKey("metadata")) {
metadata = _json["metadata"];
@@ -858,7 +849,8 @@ class Entity {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (mentions != null) {
_json["mentions"] = mentions.map((value) => (value).toJson()).toList();
}
@@ -878,20 +870,17 @@ class Entity {
}
}
-/**
- * Represents a mention for an entity in the text. Currently, proper noun
- * mentions are supported.
- */
+/// Represents a mention for an entity in the text. Currently, proper noun
+/// mentions are supported.
class EntityMention {
- /** The mention text. */
+ /// The mention text.
TextSpan text;
- /**
- * The type of the entity mention.
- * Possible string values are:
- * - "TYPE_UNKNOWN" : Unknown
- * - "PROPER" : Proper name
- * - "COMMON" : Common noun (or noun compound)
- */
+
+ /// The type of the entity mention.
+ /// Possible string values are:
+ /// - "TYPE_UNKNOWN" : Unknown
+ /// - "PROPER" : Proper name
+ /// - "COMMON" : Common noun (or noun compound)
core.String type;
EntityMention();
@@ -906,7 +895,8 @@ class EntityMention {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (text != null) {
_json["text"] = (text).toJson();
}
@@ -917,16 +907,16 @@ class EntityMention {
}
}
-/**
- * All available features for sentiment, syntax, and semantic analysis.
- * Setting each one to true will enable that specific analysis for the input.
- */
+/// All available features for sentiment, syntax, and semantic analysis.
+/// Setting each one to true will enable that specific analysis for the input.
class Features {
- /** Extract document-level sentiment. */
+ /// Extract document-level sentiment.
core.bool extractDocumentSentiment;
- /** Extract entities. */
+
+ /// Extract entities.
core.bool extractEntities;
- /** Extract syntax information. */
+
+ /// Extract syntax information.
core.bool extractSyntax;
Features();
@@ -944,7 +934,8 @@ class Features {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (extractDocumentSentiment != null) {
_json["extractDocumentSentiment"] = extractDocumentSentiment;
}
@@ -958,165 +949,150 @@ class Features {
}
}
-/**
- * Represents part of speech information for a token. Parts of speech
- * are as defined in
- * http://www.lrec-conf.org/proceedings/lrec2012/pdf/274_Paper.pdf
- */
+/// Represents part of speech information for a token. Parts of speech
+/// are as defined in
+/// http://www.lrec-conf.org/proceedings/lrec2012/pdf/274_Paper.pdf
class PartOfSpeech {
- /**
- * The grammatical aspect.
- * Possible string values are:
- * - "ASPECT_UNKNOWN" : Aspect is not applicable in the analyzed language or
- * is not predicted.
- * - "PERFECTIVE" : Perfective
- * - "IMPERFECTIVE" : Imperfective
- * - "PROGRESSIVE" : Progressive
- */
+ /// The grammatical aspect.
+ /// Possible string values are:
+ /// - "ASPECT_UNKNOWN" : Aspect is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "PERFECTIVE" : Perfective
+ /// - "IMPERFECTIVE" : Imperfective
+ /// - "PROGRESSIVE" : Progressive
core.String aspect;
- /**
- * The grammatical case.
- * Possible string values are:
- * - "CASE_UNKNOWN" : Case is not applicable in the analyzed language or is
- * not predicted.
- * - "ACCUSATIVE" : Accusative
- * - "ADVERBIAL" : Adverbial
- * - "COMPLEMENTIVE" : Complementive
- * - "DATIVE" : Dative
- * - "GENITIVE" : Genitive
- * - "INSTRUMENTAL" : Instrumental
- * - "LOCATIVE" : Locative
- * - "NOMINATIVE" : Nominative
- * - "OBLIQUE" : Oblique
- * - "PARTITIVE" : Partitive
- * - "PREPOSITIONAL" : Prepositional
- * - "REFLEXIVE_CASE" : Reflexive
- * - "RELATIVE_CASE" : Relative
- * - "VOCATIVE" : Vocative
- */
+
+ /// The grammatical case.
+ /// Possible string values are:
+ /// - "CASE_UNKNOWN" : Case is not applicable in the analyzed language or is
+ /// not predicted.
+ /// - "ACCUSATIVE" : Accusative
+ /// - "ADVERBIAL" : Adverbial
+ /// - "COMPLEMENTIVE" : Complementive
+ /// - "DATIVE" : Dative
+ /// - "GENITIVE" : Genitive
+ /// - "INSTRUMENTAL" : Instrumental
+ /// - "LOCATIVE" : Locative
+ /// - "NOMINATIVE" : Nominative
+ /// - "OBLIQUE" : Oblique
+ /// - "PARTITIVE" : Partitive
+ /// - "PREPOSITIONAL" : Prepositional
+ /// - "REFLEXIVE_CASE" : Reflexive
+ /// - "RELATIVE_CASE" : Relative
+ /// - "VOCATIVE" : Vocative
core.String case_;
- /**
- * The grammatical form.
- * Possible string values are:
- * - "FORM_UNKNOWN" : Form is not applicable in the analyzed language or is
- * not predicted.
- * - "ADNOMIAL" : Adnomial
- * - "AUXILIARY" : Auxiliary
- * - "COMPLEMENTIZER" : Complementizer
- * - "FINAL_ENDING" : Final ending
- * - "GERUND" : Gerund
- * - "REALIS" : Realis
- * - "IRREALIS" : Irrealis
- * - "SHORT" : Short form
- * - "LONG" : Long form
- * - "ORDER" : Order form
- * - "SPECIFIC" : Specific form
- */
+
+ /// The grammatical form.
+ /// Possible string values are:
+ /// - "FORM_UNKNOWN" : Form is not applicable in the analyzed language or is
+ /// not predicted.
+ /// - "ADNOMIAL" : Adnomial
+ /// - "AUXILIARY" : Auxiliary
+ /// - "COMPLEMENTIZER" : Complementizer
+ /// - "FINAL_ENDING" : Final ending
+ /// - "GERUND" : Gerund
+ /// - "REALIS" : Realis
+ /// - "IRREALIS" : Irrealis
+ /// - "SHORT" : Short form
+ /// - "LONG" : Long form
+ /// - "ORDER" : Order form
+ /// - "SPECIFIC" : Specific form
core.String form;
- /**
- * The grammatical gender.
- * Possible string values are:
- * - "GENDER_UNKNOWN" : Gender is not applicable in the analyzed language or
- * is not predicted.
- * - "FEMININE" : Feminine
- * - "MASCULINE" : Masculine
- * - "NEUTER" : Neuter
- */
+
+ /// The grammatical gender.
+ /// Possible string values are:
+ /// - "GENDER_UNKNOWN" : Gender is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "FEMININE" : Feminine
+ /// - "MASCULINE" : Masculine
+ /// - "NEUTER" : Neuter
core.String gender;
- /**
- * The grammatical mood.
- * Possible string values are:
- * - "MOOD_UNKNOWN" : Mood is not applicable in the analyzed language or is
- * not predicted.
- * - "CONDITIONAL_MOOD" : Conditional
- * - "IMPERATIVE" : Imperative
- * - "INDICATIVE" : Indicative
- * - "INTERROGATIVE" : Interrogative
- * - "JUSSIVE" : Jussive
- * - "SUBJUNCTIVE" : Subjunctive
- */
+
+ /// The grammatical mood.
+ /// Possible string values are:
+ /// - "MOOD_UNKNOWN" : Mood is not applicable in the analyzed language or is
+ /// not predicted.
+ /// - "CONDITIONAL_MOOD" : Conditional
+ /// - "IMPERATIVE" : Imperative
+ /// - "INDICATIVE" : Indicative
+ /// - "INTERROGATIVE" : Interrogative
+ /// - "JUSSIVE" : Jussive
+ /// - "SUBJUNCTIVE" : Subjunctive
core.String mood;
- /**
- * The grammatical number.
- * Possible string values are:
- * - "NUMBER_UNKNOWN" : Number is not applicable in the analyzed language or
- * is not predicted.
- * - "SINGULAR" : Singular
- * - "PLURAL" : Plural
- * - "DUAL" : Dual
- */
+
+ /// The grammatical number.
+ /// Possible string values are:
+ /// - "NUMBER_UNKNOWN" : Number is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "SINGULAR" : Singular
+ /// - "PLURAL" : Plural
+ /// - "DUAL" : Dual
core.String number;
- /**
- * The grammatical person.
- * Possible string values are:
- * - "PERSON_UNKNOWN" : Person is not applicable in the analyzed language or
- * is not predicted.
- * - "FIRST" : First
- * - "SECOND" : Second
- * - "THIRD" : Third
- * - "REFLEXIVE_PERSON" : Reflexive
- */
+
+ /// The grammatical person.
+ /// Possible string values are:
+ /// - "PERSON_UNKNOWN" : Person is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "FIRST" : First
+ /// - "SECOND" : Second
+ /// - "THIRD" : Third
+ /// - "REFLEXIVE_PERSON" : Reflexive
core.String person;
- /**
- * The grammatical properness.
- * Possible string values are:
- * - "PROPER_UNKNOWN" : Proper is not applicable in the analyzed language or
- * is not predicted.
- * - "PROPER" : Proper
- * - "NOT_PROPER" : Not proper
- */
+
+ /// The grammatical properness.
+ /// Possible string values are:
+ /// - "PROPER_UNKNOWN" : Proper is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "PROPER" : Proper
+ /// - "NOT_PROPER" : Not proper
core.String proper;
- /**
- * The grammatical reciprocity.
- * Possible string values are:
- * - "RECIPROCITY_UNKNOWN" : Reciprocity is not applicable in the analyzed
- * language or is not
- * predicted.
- * - "RECIPROCAL" : Reciprocal
- * - "NON_RECIPROCAL" : Non-reciprocal
- */
+
+ /// The grammatical reciprocity.
+ /// Possible string values are:
+ /// - "RECIPROCITY_UNKNOWN" : Reciprocity is not applicable in the analyzed
+ /// language or is not
+ /// predicted.
+ /// - "RECIPROCAL" : Reciprocal
+ /// - "NON_RECIPROCAL" : Non-reciprocal
core.String reciprocity;
- /**
- * The part of speech tag.
- * Possible string values are:
- * - "UNKNOWN" : Unknown
- * - "ADJ" : Adjective
- * - "ADP" : Adposition (preposition and postposition)
- * - "ADV" : Adverb
- * - "CONJ" : Conjunction
- * - "DET" : Determiner
- * - "NOUN" : Noun (common and proper)
- * - "NUM" : Cardinal number
- * - "PRON" : Pronoun
- * - "PRT" : Particle or other function word
- * - "PUNCT" : Punctuation
- * - "VERB" : Verb (all tenses and modes)
- * - "X" : Other: foreign words, typos, abbreviations
- * - "AFFIX" : Affix
- */
+
+ /// The part of speech tag.
+ /// Possible string values are:
+ /// - "UNKNOWN" : Unknown
+ /// - "ADJ" : Adjective
+ /// - "ADP" : Adposition (preposition and postposition)
+ /// - "ADV" : Adverb
+ /// - "CONJ" : Conjunction
+ /// - "DET" : Determiner
+ /// - "NOUN" : Noun (common and proper)
+ /// - "NUM" : Cardinal number
+ /// - "PRON" : Pronoun
+ /// - "PRT" : Particle or other function word
+ /// - "PUNCT" : Punctuation
+ /// - "VERB" : Verb (all tenses and modes)
+ /// - "X" : Other: foreign words, typos, abbreviations
+ /// - "AFFIX" : Affix
core.String tag;
- /**
- * The grammatical tense.
- * Possible string values are:
- * - "TENSE_UNKNOWN" : Tense is not applicable in the analyzed language or is
- * not predicted.
- * - "CONDITIONAL_TENSE" : Conditional
- * - "FUTURE" : Future
- * - "PAST" : Past
- * - "PRESENT" : Present
- * - "IMPERFECT" : Imperfect
- * - "PLUPERFECT" : Pluperfect
- */
+
+ /// The grammatical tense.
+ /// Possible string values are:
+ /// - "TENSE_UNKNOWN" : Tense is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "CONDITIONAL_TENSE" : Conditional
+ /// - "FUTURE" : Future
+ /// - "PAST" : Past
+ /// - "PRESENT" : Present
+ /// - "IMPERFECT" : Imperfect
+ /// - "PLUPERFECT" : Pluperfect
core.String tense;
- /**
- * The grammatical voice.
- * Possible string values are:
- * - "VOICE_UNKNOWN" : Voice is not applicable in the analyzed language or is
- * not predicted.
- * - "ACTIVE" : Active
- * - "CAUSATIVE" : Causative
- * - "PASSIVE" : Passive
- */
+
+ /// The grammatical voice.
+ /// Possible string values are:
+ /// - "VOICE_UNKNOWN" : Voice is not applicable in the analyzed language or
+ /// is not predicted.
+ /// - "ACTIVE" : Active
+ /// - "CAUSATIVE" : Causative
+ /// - "PASSIVE" : Passive
core.String voice;
PartOfSpeech();
@@ -1161,7 +1137,8 @@ class PartOfSpeech {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (aspect != null) {
_json["aspect"] = aspect;
}
@@ -1202,15 +1179,14 @@ class PartOfSpeech {
}
}
-/** Represents a sentence in the input document. */
+/// Represents a sentence in the input document.
class Sentence {
- /**
- * For calls to AnalyzeSentiment or if
- * AnnotateTextRequest.Features.extract_document_sentiment is set to
- * true, this field will contain the sentiment for the sentence.
- */
+ /// For calls to AnalyzeSentiment or if
+ /// AnnotateTextRequest.Features.extract_document_sentiment is set to
+ /// true, this field will contain the sentiment for the sentence.
Sentiment sentiment;
- /** The sentence text. */
+
+ /// The sentence text.
TextSpan text;
Sentence();
@@ -1225,7 +1201,8 @@ class Sentence {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (sentiment != null) {
_json["sentiment"] = (sentiment).toJson();
}
@@ -1236,21 +1213,16 @@ class Sentence {
}
}
-/**
- * Represents the feeling associated with the entire text or entities in
- * the text.
- */
+/// Represents the feeling associated with the entire text or entities in
+/// the text.
class Sentiment {
- /**
- * A non-negative number in the [0, +inf) range, which represents
- * the absolute magnitude of sentiment regardless of score (positive or
- * negative).
- */
+ /// A non-negative number in the [0, +inf) range, which represents
+ /// the absolute magnitude of sentiment regardless of score (positive or
+ /// negative).
core.double magnitude;
- /**
- * Sentiment score between -1.0 (negative sentiment) and 1.0
- * (positive sentiment).
- */
+
+ /// Sentiment score between -1.0 (negative sentiment) and 1.0
+ /// (positive sentiment).
core.double score;
Sentiment();
@@ -1265,7 +1237,8 @@ class Sentiment {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (magnitude != null) {
_json["magnitude"] = magnitude;
}
@@ -1276,78 +1249,81 @@ class Sentiment {
}
}
-/**
- * The `Status` type defines a logical error model that is suitable for
- * different
- * programming environments, including REST APIs and RPC APIs. It is used by
- * [gRPC](https://github.com/grpc). The error model is designed to be:
- *
- * - Simple to use and understand for most users
- * - Flexible enough to meet unexpected needs
- *
- * # Overview
- *
- * The `Status` message contains three pieces of data: error code, error
- * message,
- * and error details. The error code should be an enum value of
- * google.rpc.Code, but it may accept additional error codes if needed. The
- * error message should be a developer-facing English message that helps
- * developers *understand* and *resolve* the error. If a localized user-facing
- * error message is needed, put the localized message in the error details or
- * localize it in the client. The optional error details may contain arbitrary
- * information about the error. There is a predefined set of error detail types
- * in the package `google.rpc` that can be used for common error conditions.
- *
- * # Language mapping
- *
- * The `Status` message is the logical representation of the error model, but it
- * is not necessarily the actual wire format. When the `Status` message is
- * exposed in different client libraries and different wire protocols, it can be
- * mapped differently. For example, it will likely be mapped to some exceptions
- * in Java, but more likely mapped to some error codes in C.
- *
- * # Other uses
- *
- * The error model and the `Status` message can be used in a variety of
- * environments, either with or without APIs, to provide a
- * consistent developer experience across different environments.
- *
- * Example uses of this error model include:
- *
- * - Partial errors. If a service needs to return partial errors to the client,
- * it may embed the `Status` in the normal response to indicate the partial
- * errors.
- *
- * - Workflow errors. A typical workflow has multiple steps. Each step may
- * have a `Status` message for error reporting.
- *
- * - Batch operations. If a client uses batch request and batch response, the
- * `Status` message should be used directly inside batch response, one for
- * each error sub-response.
- *
- * - Asynchronous operations. If an API call embeds asynchronous operation
- * results in its response, the status of those operations should be
- * represented directly using the `Status` message.
- *
- * - Logging. If some API errors are stored in logs, the message `Status` could
- * be used directly after any stripping needed for security/privacy reasons.
- */
+/// The `Status` type defines a logical error model that is suitable for
+/// different
+/// programming environments, including REST APIs and RPC APIs. It is used by
+/// [gRPC](https://github.com/grpc). The error model is designed to be:
+///
+/// - Simple to use and understand for most users
+/// - Flexible enough to meet unexpected needs
+///
+/// # Overview
+///
+/// The `Status` message contains three pieces of data: error code, error
+/// message,
+/// and error details. The error code should be an enum value of
+/// google.rpc.Code, but it may accept additional error codes if needed. The
+/// error message should be a developer-facing English message that helps
+/// developers *understand* and *resolve* the error. If a localized user-facing
+/// error message is needed, put the localized message in the error details or
+/// localize it in the client. The optional error details may contain arbitrary
+/// information about the error. There is a predefined set of error detail
+/// types
+/// in the package `google.rpc` that can be used for common error conditions.
+///
+/// # Language mapping
+///
+/// The `Status` message is the logical representation of the error model, but
+/// it
+/// is not necessarily the actual wire format. When the `Status` message is
+/// exposed in different client libraries and different wire protocols, it can
+/// be
+/// mapped differently. For example, it will likely be mapped to some
+/// exceptions
+/// in Java, but more likely mapped to some error codes in C.
+///
+/// # Other uses
+///
+/// The error model and the `Status` message can be used in a variety of
+/// environments, either with or without APIs, to provide a
+/// consistent developer experience across different environments.
+///
+/// Example uses of this error model include:
+///
+/// - Partial errors. If a service needs to return partial errors to the
+/// client,
+/// it may embed the `Status` in the normal response to indicate the partial
+/// errors.
+///
+/// - Workflow errors. A typical workflow has multiple steps. Each step may
+/// have a `Status` message for error reporting.
+///
+/// - Batch operations. If a client uses batch request and batch response, the
+/// `Status` message should be used directly inside batch response, one for
+/// each error sub-response.
+///
+/// - Asynchronous operations. If an API call embeds asynchronous operation
+/// results in its response, the status of those operations should be
+/// represented directly using the `Status` message.
+///
+/// - Logging. If some API errors are stored in logs, the message `Status`
+/// could
+/// be used directly after any stripping needed for security/privacy reasons.
class Status {
- /** The status code, which should be an enum value of google.rpc.Code. */
+ /// The status code, which should be an enum value of google.rpc.Code.
core.int code;
- /**
- * A list of messages that carry the error details. There is a common set of
- * message types for APIs to use.
- *
- * The values for Object must be JSON objects. It can consist of `num`,
- * `String`, `bool` and `null` as well as `Map` and `List` values.
- */
+
+ /// A list of messages that carry the error details. There is a common set
+ /// of
+ /// message types for APIs to use.
+ ///
+ /// The values for Object must be JSON objects. It can consist of `num`,
+ /// `String`, `bool` and `null` as well as `Map` and `List` values.
core.List<core.Map<core.String, core.Object>> details;
- /**
- * A developer-facing error message, which should be in English. Any
- * user-facing error message should be localized and sent in the
- * google.rpc.Status.details field, or localized by the client.
- */
+
+ /// A developer-facing error message, which should be in English. Any
+ /// user-facing error message should be localized and sent in the
+ /// google.rpc.Status.details field, or localized by the client.
core.String message;
Status();
@@ -1365,7 +1341,8 @@ class Status {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (code != null) {
_json["code"] = code;
}
@@ -1379,14 +1356,13 @@ class Status {
}
}
-/** Represents an output piece of text. */
+/// Represents an output piece of text.
class TextSpan {
- /**
- * The API calculates the beginning offset of the content in the original
- * document according to the EncodingType specified in the API request.
- */
+ /// The API calculates the beginning offset of the content in the original
+ /// document according to the EncodingType specified in the API request.
core.int beginOffset;
- /** The content of the output text. */
+
+ /// The content of the output text.
core.String content;
TextSpan();
@@ -1401,7 +1377,8 @@ class TextSpan {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (beginOffset != null) {
_json["beginOffset"] = beginOffset;
}
@@ -1412,17 +1389,19 @@ class TextSpan {
}
}
-/** Represents the smallest syntactic building block of the text. */
+/// Represents the smallest syntactic building block of the text.
class Token {
- /** Dependency tree parse for this token. */
+ /// Dependency tree parse for this token.
DependencyEdge dependencyEdge;
- /**
- * [Lemma](https://en.wikipedia.org/wiki/Lemma_%28morphology%29) of the token.
- */
+
+ /// [Lemma](https://en.wikipedia.org/wiki/Lemma_%28morphology%29) of the
+ /// token.
core.String lemma;
- /** Parts of speech tag for this token. */
+
+ /// Parts of speech tag for this token.
PartOfSpeech partOfSpeech;
- /** The token text. */
+
+ /// The token text.
TextSpan text;
Token();
@@ -1443,7 +1422,8 @@ class Token {
}
core.Map<core.String, core.Object> toJson() {
- final core.Map<core.String, core.Object> _json = new core.Map<core.String, core.Object>();
+ final core.Map<core.String, core.Object> _json =
+ new core.Map<core.String, core.Object>();
if (dependencyEdge != null) {
_json["dependencyEdge"] = (dependencyEdge).toJson();
}
« no previous file with comments | « generated/googleapis/lib/kgsearch/v1.dart ('k') | generated/googleapis/lib/licensing/v1.dart » ('j') | no next file with comments »

Powered by Google App Engine
This is Rietveld 408576698