Index: third_party/protobuf/conformance/conformance_test.cc |
diff --git a/third_party/protobuf/conformance/conformance_test.cc b/third_party/protobuf/conformance/conformance_test.cc |
new file mode 100644 |
index 0000000000000000000000000000000000000000..135e82dc4a0094660d2b5a53277b38339674f2bf |
--- /dev/null |
+++ b/third_party/protobuf/conformance/conformance_test.cc |
@@ -0,0 +1,1992 @@ |
+// Protocol Buffers - Google's data interchange format |
+// Copyright 2008 Google Inc. All rights reserved. |
+// https://developers.google.com/protocol-buffers/ |
+// |
+// Redistribution and use in source and binary forms, with or without |
+// modification, are permitted provided that the following conditions are |
+// met: |
+// |
+// * Redistributions of source code must retain the above copyright |
+// notice, this list of conditions and the following disclaimer. |
+// * Redistributions in binary form must reproduce the above |
+// copyright notice, this list of conditions and the following disclaimer |
+// in the documentation and/or other materials provided with the |
+// distribution. |
+// * Neither the name of Google Inc. nor the names of its |
+// contributors may be used to endorse or promote products derived from |
+// this software without specific prior written permission. |
+// |
+// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
+// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
+// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
+// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
+// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
+// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
+// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
+// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
+// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
+// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
+// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
+ |
+#include <stdarg.h> |
+#include <string> |
+ |
+#include "conformance.pb.h" |
+#include "conformance_test.h" |
+#include <google/protobuf/stubs/common.h> |
+#include <google/protobuf/stubs/stringprintf.h> |
+#include <google/protobuf/text_format.h> |
+#include <google/protobuf/util/json_util.h> |
+#include <google/protobuf/util/field_comparator.h> |
+#include <google/protobuf/util/message_differencer.h> |
+#include <google/protobuf/util/type_resolver_util.h> |
+#include <google/protobuf/wire_format_lite.h> |
+ |
+#include "third_party/jsoncpp/json.h" |
+ |
+using conformance::ConformanceRequest; |
+using conformance::ConformanceResponse; |
+using conformance::TestAllTypes; |
+using conformance::WireFormat; |
+using google::protobuf::Descriptor; |
+using google::protobuf::FieldDescriptor; |
+using google::protobuf::internal::WireFormatLite; |
+using google::protobuf::TextFormat; |
+using google::protobuf::util::DefaultFieldComparator; |
+using google::protobuf::util::JsonToBinaryString; |
+using google::protobuf::util::MessageDifferencer; |
+using google::protobuf::util::NewTypeResolverForDescriptorPool; |
+using google::protobuf::util::Status; |
+using std::string; |
+ |
+namespace { |
+ |
+static const char kTypeUrlPrefix[] = "type.googleapis.com"; |
+ |
+static string GetTypeUrl(const Descriptor* message) { |
+ return string(kTypeUrlPrefix) + "/" + message->full_name(); |
+} |
+ |
+/* Routines for building arbitrary protos *************************************/ |
+ |
+// We would use CodedOutputStream except that we want more freedom to build |
+// arbitrary protos (even invalid ones). |
+ |
+const string empty; |
+ |
+string cat(const string& a, const string& b, |
+ const string& c = empty, |
+ const string& d = empty, |
+ const string& e = empty, |
+ const string& f = empty, |
+ const string& g = empty, |
+ const string& h = empty, |
+ const string& i = empty, |
+ const string& j = empty, |
+ const string& k = empty, |
+ const string& l = empty) { |
+ string ret; |
+ ret.reserve(a.size() + b.size() + c.size() + d.size() + e.size() + f.size() + |
+ g.size() + h.size() + i.size() + j.size() + k.size() + l.size()); |
+ ret.append(a); |
+ ret.append(b); |
+ ret.append(c); |
+ ret.append(d); |
+ ret.append(e); |
+ ret.append(f); |
+ ret.append(g); |
+ ret.append(h); |
+ ret.append(i); |
+ ret.append(j); |
+ ret.append(k); |
+ ret.append(l); |
+ return ret; |
+} |
+ |
+// The maximum number of bytes that it takes to encode a 64-bit varint. |
+#define VARINT_MAX_LEN 10 |
+ |
+size_t vencode64(uint64_t val, char *buf) { |
+ if (val == 0) { buf[0] = 0; return 1; } |
+ size_t i = 0; |
+ while (val) { |
+ uint8_t byte = val & 0x7fU; |
+ val >>= 7; |
+ if (val) byte |= 0x80U; |
+ buf[i++] = byte; |
+ } |
+ return i; |
+} |
+ |
+string varint(uint64_t x) { |
+ char buf[VARINT_MAX_LEN]; |
+ size_t len = vencode64(x, buf); |
+ return string(buf, len); |
+} |
+ |
+// TODO: proper byte-swapping for big-endian machines. |
+string fixed32(void *data) { return string(static_cast<char*>(data), 4); } |
+string fixed64(void *data) { return string(static_cast<char*>(data), 8); } |
+ |
+string delim(const string& buf) { return cat(varint(buf.size()), buf); } |
+string uint32(uint32_t u32) { return fixed32(&u32); } |
+string uint64(uint64_t u64) { return fixed64(&u64); } |
+string flt(float f) { return fixed32(&f); } |
+string dbl(double d) { return fixed64(&d); } |
+string zz32(int32_t x) { return varint(WireFormatLite::ZigZagEncode32(x)); } |
+string zz64(int64_t x) { return varint(WireFormatLite::ZigZagEncode64(x)); } |
+ |
+string tag(uint32_t fieldnum, char wire_type) { |
+ return varint((fieldnum << 3) | wire_type); |
+} |
+ |
+string submsg(uint32_t fn, const string& buf) { |
+ return cat( tag(fn, WireFormatLite::WIRETYPE_LENGTH_DELIMITED), delim(buf) ); |
+} |
+ |
+#define UNKNOWN_FIELD 666 |
+ |
+uint32_t GetFieldNumberForType(FieldDescriptor::Type type, bool repeated) { |
+ const Descriptor* d = TestAllTypes().GetDescriptor(); |
+ for (int i = 0; i < d->field_count(); i++) { |
+ const FieldDescriptor* f = d->field(i); |
+ if (f->type() == type && f->is_repeated() == repeated) { |
+ return f->number(); |
+ } |
+ } |
+ GOOGLE_LOG(FATAL) << "Couldn't find field with type " << (int)type; |
+ return 0; |
+} |
+ |
+string UpperCase(string str) { |
+ for (int i = 0; i < str.size(); i++) { |
+ str[i] = toupper(str[i]); |
+ } |
+ return str; |
+} |
+ |
+} // anonymous namespace |
+ |
+namespace google { |
+namespace protobuf { |
+ |
+void ConformanceTestSuite::ReportSuccess(const string& test_name) { |
+ if (expected_to_fail_.erase(test_name) != 0) { |
+ StringAppendF(&output_, |
+ "ERROR: test %s is in the failure list, but test succeeded. " |
+ "Remove it from the failure list.\n", |
+ test_name.c_str()); |
+ unexpected_succeeding_tests_.insert(test_name); |
+ } |
+ successes_++; |
+} |
+ |
+void ConformanceTestSuite::ReportFailure(const string& test_name, |
+ const ConformanceRequest& request, |
+ const ConformanceResponse& response, |
+ const char* fmt, ...) { |
+ if (expected_to_fail_.erase(test_name) == 1) { |
+ expected_failures_++; |
+ if (!verbose_) |
+ return; |
+ } else { |
+ StringAppendF(&output_, "ERROR, test=%s: ", test_name.c_str()); |
+ unexpected_failing_tests_.insert(test_name); |
+ } |
+ va_list args; |
+ va_start(args, fmt); |
+ StringAppendV(&output_, fmt, args); |
+ va_end(args); |
+ StringAppendF(&output_, " request=%s, response=%s\n", |
+ request.ShortDebugString().c_str(), |
+ response.ShortDebugString().c_str()); |
+} |
+ |
+void ConformanceTestSuite::ReportSkip(const string& test_name, |
+ const ConformanceRequest& request, |
+ const ConformanceResponse& response) { |
+ if (verbose_) { |
+ StringAppendF(&output_, "SKIPPED, test=%s request=%s, response=%s\n", |
+ test_name.c_str(), request.ShortDebugString().c_str(), |
+ response.ShortDebugString().c_str()); |
+ } |
+ skipped_.insert(test_name); |
+} |
+ |
+void ConformanceTestSuite::RunTest(const string& test_name, |
+ const ConformanceRequest& request, |
+ ConformanceResponse* response) { |
+ if (test_names_.insert(test_name).second == false) { |
+ GOOGLE_LOG(FATAL) << "Duplicated test name: " << test_name; |
+ } |
+ |
+ string serialized_request; |
+ string serialized_response; |
+ request.SerializeToString(&serialized_request); |
+ |
+ runner_->RunTest(test_name, serialized_request, &serialized_response); |
+ |
+ if (!response->ParseFromString(serialized_response)) { |
+ response->Clear(); |
+ response->set_runtime_error("response proto could not be parsed."); |
+ } |
+ |
+ if (verbose_) { |
+ StringAppendF(&output_, "conformance test: name=%s, request=%s, response=%s\n", |
+ test_name.c_str(), |
+ request.ShortDebugString().c_str(), |
+ response->ShortDebugString().c_str()); |
+ } |
+} |
+ |
+void ConformanceTestSuite::RunValidInputTest( |
+ const string& test_name, const string& input, WireFormat input_format, |
+ const string& equivalent_text_format, WireFormat requested_output) { |
+ TestAllTypes reference_message; |
+ GOOGLE_CHECK( |
+ TextFormat::ParseFromString(equivalent_text_format, &reference_message)) |
+ << "Failed to parse data for test case: " << test_name |
+ << ", data: " << equivalent_text_format; |
+ |
+ ConformanceRequest request; |
+ ConformanceResponse response; |
+ |
+ switch (input_format) { |
+ case conformance::PROTOBUF: |
+ request.set_protobuf_payload(input); |
+ break; |
+ |
+ case conformance::JSON: |
+ request.set_json_payload(input); |
+ break; |
+ |
+ default: |
+ GOOGLE_LOG(FATAL) << "Unspecified input format"; |
+ } |
+ |
+ request.set_requested_output_format(requested_output); |
+ |
+ RunTest(test_name, request, &response); |
+ |
+ TestAllTypes test_message; |
+ |
+ switch (response.result_case()) { |
+ case ConformanceResponse::kParseError: |
+ case ConformanceResponse::kRuntimeError: |
+ case ConformanceResponse::kSerializeError: |
+ ReportFailure(test_name, request, response, |
+ "Failed to parse JSON input or produce JSON output."); |
+ return; |
+ |
+ case ConformanceResponse::kSkipped: |
+ ReportSkip(test_name, request, response); |
+ return; |
+ |
+ case ConformanceResponse::kJsonPayload: { |
+ if (requested_output != conformance::JSON) { |
+ ReportFailure( |
+ test_name, request, response, |
+ "Test was asked for protobuf output but provided JSON instead."); |
+ return; |
+ } |
+ string binary_protobuf; |
+ Status status = |
+ JsonToBinaryString(type_resolver_.get(), type_url_, |
+ response.json_payload(), &binary_protobuf); |
+ if (!status.ok()) { |
+ ReportFailure(test_name, request, response, |
+ "JSON output we received from test was unparseable."); |
+ return; |
+ } |
+ |
+ GOOGLE_CHECK(test_message.ParseFromString(binary_protobuf)); |
+ break; |
+ } |
+ |
+ case ConformanceResponse::kProtobufPayload: { |
+ if (requested_output != conformance::PROTOBUF) { |
+ ReportFailure( |
+ test_name, request, response, |
+ "Test was asked for JSON output but provided protobuf instead."); |
+ return; |
+ } |
+ |
+ if (!test_message.ParseFromString(response.protobuf_payload())) { |
+ ReportFailure(test_name, request, response, |
+ "Protobuf output we received from test was unparseable."); |
+ return; |
+ } |
+ |
+ break; |
+ } |
+ |
+ default: |
+ GOOGLE_LOG(FATAL) << test_name << ": unknown payload type: " |
+ << response.result_case(); |
+ } |
+ |
+ MessageDifferencer differencer; |
+ DefaultFieldComparator field_comparator; |
+ field_comparator.set_treat_nan_as_equal(true); |
+ differencer.set_field_comparator(&field_comparator); |
+ string differences; |
+ differencer.ReportDifferencesToString(&differences); |
+ |
+ if (differencer.Compare(reference_message, test_message)) { |
+ ReportSuccess(test_name); |
+ } else { |
+ ReportFailure(test_name, request, response, |
+ "Output was not equivalent to reference message: %s.", |
+ differences.c_str()); |
+ } |
+} |
+ |
+// Expect that this precise protobuf will cause a parse error. |
+void ConformanceTestSuite::ExpectParseFailureForProto( |
+ const string& proto, const string& test_name) { |
+ ConformanceRequest request; |
+ ConformanceResponse response; |
+ request.set_protobuf_payload(proto); |
+ string effective_test_name = "ProtobufInput." + test_name; |
+ |
+ // We don't expect output, but if the program erroneously accepts the protobuf |
+ // we let it send its response as this. We must not leave it unspecified. |
+ request.set_requested_output_format(conformance::PROTOBUF); |
+ |
+ RunTest(effective_test_name, request, &response); |
+ if (response.result_case() == ConformanceResponse::kParseError) { |
+ ReportSuccess(effective_test_name); |
+ } else { |
+ ReportFailure(effective_test_name, request, response, |
+ "Should have failed to parse, but didn't."); |
+ } |
+} |
+ |
+// Expect that this protobuf will cause a parse error, even if it is followed |
+// by valid protobuf data. We can try running this twice: once with this |
+// data verbatim and once with this data followed by some valid data. |
+// |
+// TODO(haberman): implement the second of these. |
+void ConformanceTestSuite::ExpectHardParseFailureForProto( |
+ const string& proto, const string& test_name) { |
+ return ExpectParseFailureForProto(proto, test_name); |
+} |
+ |
+void ConformanceTestSuite::RunValidJsonTest( |
+ const string& test_name, const string& input_json, |
+ const string& equivalent_text_format) { |
+ RunValidInputTest("JsonInput." + test_name + ".ProtobufOutput", input_json, |
+ conformance::JSON, equivalent_text_format, |
+ conformance::PROTOBUF); |
+ RunValidInputTest("JsonInput." + test_name + ".JsonOutput", input_json, |
+ conformance::JSON, equivalent_text_format, |
+ conformance::JSON); |
+} |
+ |
+void ConformanceTestSuite::RunValidJsonTestWithProtobufInput( |
+ const string& test_name, const TestAllTypes& input, |
+ const string& equivalent_text_format) { |
+ RunValidInputTest("ProtobufInput." + test_name + ".JsonOutput", |
+ input.SerializeAsString(), conformance::PROTOBUF, |
+ equivalent_text_format, conformance::JSON); |
+} |
+ |
+// According to proto3 JSON specification, JSON serializers follow more strict |
+// rules than parsers (e.g., a serializer must serialize int32 values as JSON |
+// numbers while the parser is allowed to accept them as JSON strings). This |
+// method allows strict checking on a proto3 JSON serializer by inspecting |
+// the JSON output directly. |
+void ConformanceTestSuite::RunValidJsonTestWithValidator( |
+ const string& test_name, const string& input_json, |
+ const Validator& validator) { |
+ ConformanceRequest request; |
+ ConformanceResponse response; |
+ request.set_json_payload(input_json); |
+ request.set_requested_output_format(conformance::JSON); |
+ |
+ string effective_test_name = "JsonInput." + test_name + ".Validator"; |
+ |
+ RunTest(effective_test_name, request, &response); |
+ |
+ if (response.result_case() != ConformanceResponse::kJsonPayload) { |
+ ReportFailure(effective_test_name, request, response, |
+ "Expected JSON payload but got type %d.", |
+ response.result_case()); |
+ return; |
+ } |
+ Json::Reader reader; |
+ Json::Value value; |
+ if (!reader.parse(response.json_payload(), value)) { |
+ ReportFailure(effective_test_name, request, response, |
+ "JSON payload cannot be parsed as valid JSON: %s", |
+ reader.getFormattedErrorMessages().c_str()); |
+ return; |
+ } |
+ if (!validator(value)) { |
+ ReportFailure(effective_test_name, request, response, |
+ "JSON payload validation failed."); |
+ return; |
+ } |
+ ReportSuccess(effective_test_name); |
+} |
+ |
+void ConformanceTestSuite::ExpectParseFailureForJson( |
+ const string& test_name, const string& input_json) { |
+ ConformanceRequest request; |
+ ConformanceResponse response; |
+ request.set_json_payload(input_json); |
+ string effective_test_name = "JsonInput." + test_name; |
+ |
+ // We don't expect output, but if the program erroneously accepts the protobuf |
+ // we let it send its response as this. We must not leave it unspecified. |
+ request.set_requested_output_format(conformance::JSON); |
+ |
+ RunTest(effective_test_name, request, &response); |
+ if (response.result_case() == ConformanceResponse::kParseError) { |
+ ReportSuccess(effective_test_name); |
+ } else { |
+ ReportFailure(effective_test_name, request, response, |
+ "Should have failed to parse, but didn't."); |
+ } |
+} |
+ |
+void ConformanceTestSuite::ExpectSerializeFailureForJson( |
+ const string& test_name, const string& text_format) { |
+ TestAllTypes payload_message; |
+ GOOGLE_CHECK( |
+ TextFormat::ParseFromString(text_format, &payload_message)) |
+ << "Failed to parse: " << text_format; |
+ |
+ ConformanceRequest request; |
+ ConformanceResponse response; |
+ request.set_protobuf_payload(payload_message.SerializeAsString()); |
+ string effective_test_name = test_name + ".JsonOutput"; |
+ request.set_requested_output_format(conformance::JSON); |
+ |
+ RunTest(effective_test_name, request, &response); |
+ if (response.result_case() == ConformanceResponse::kSerializeError) { |
+ ReportSuccess(effective_test_name); |
+ } else { |
+ ReportFailure(effective_test_name, request, response, |
+ "Should have failed to serialize, but didn't."); |
+ } |
+} |
+ |
+void ConformanceTestSuite::TestPrematureEOFForType(FieldDescriptor::Type type) { |
+ // Incomplete values for each wire type. |
+ static const string incompletes[6] = { |
+ string("\x80"), // VARINT |
+ string("abcdefg"), // 64BIT |
+ string("\x80"), // DELIMITED (partial length) |
+ string(), // START_GROUP (no value required) |
+ string(), // END_GROUP (no value required) |
+ string("abc") // 32BIT |
+ }; |
+ |
+ uint32_t fieldnum = GetFieldNumberForType(type, false); |
+ uint32_t rep_fieldnum = GetFieldNumberForType(type, true); |
+ WireFormatLite::WireType wire_type = WireFormatLite::WireTypeForFieldType( |
+ static_cast<WireFormatLite::FieldType>(type)); |
+ const string& incomplete = incompletes[wire_type]; |
+ const string type_name = |
+ UpperCase(string(".") + FieldDescriptor::TypeName(type)); |
+ |
+ ExpectParseFailureForProto( |
+ tag(fieldnum, wire_type), |
+ "PrematureEofBeforeKnownNonRepeatedValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ tag(rep_fieldnum, wire_type), |
+ "PrematureEofBeforeKnownRepeatedValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ tag(UNKNOWN_FIELD, wire_type), |
+ "PrematureEofBeforeUnknownValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ cat( tag(fieldnum, wire_type), incomplete ), |
+ "PrematureEofInsideKnownNonRepeatedValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ cat( tag(rep_fieldnum, wire_type), incomplete ), |
+ "PrematureEofInsideKnownRepeatedValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ cat( tag(UNKNOWN_FIELD, wire_type), incomplete ), |
+ "PrematureEofInsideUnknownValue" + type_name); |
+ |
+ if (wire_type == WireFormatLite::WIRETYPE_LENGTH_DELIMITED) { |
+ ExpectParseFailureForProto( |
+ cat( tag(fieldnum, wire_type), varint(1) ), |
+ "PrematureEofInDelimitedDataForKnownNonRepeatedValue" + type_name); |
+ |
+ ExpectParseFailureForProto( |
+ cat( tag(rep_fieldnum, wire_type), varint(1) ), |
+ "PrematureEofInDelimitedDataForKnownRepeatedValue" + type_name); |
+ |
+ // EOF in the middle of delimited data for unknown value. |
+ ExpectParseFailureForProto( |
+ cat( tag(UNKNOWN_FIELD, wire_type), varint(1) ), |
+ "PrematureEofInDelimitedDataForUnknownValue" + type_name); |
+ |
+ if (type == FieldDescriptor::TYPE_MESSAGE) { |
+ // Submessage ends in the middle of a value. |
+ string incomplete_submsg = |
+ cat( tag(WireFormatLite::TYPE_INT32, WireFormatLite::WIRETYPE_VARINT), |
+ incompletes[WireFormatLite::WIRETYPE_VARINT] ); |
+ ExpectHardParseFailureForProto( |
+ cat( tag(fieldnum, WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
+ varint(incomplete_submsg.size()), |
+ incomplete_submsg ), |
+ "PrematureEofInSubmessageValue" + type_name); |
+ } |
+ } else if (type != FieldDescriptor::TYPE_GROUP) { |
+ // Non-delimited, non-group: eligible for packing. |
+ |
+ // Packed region ends in the middle of a value. |
+ ExpectHardParseFailureForProto( |
+ cat( tag(rep_fieldnum, WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
+ varint(incomplete.size()), |
+ incomplete ), |
+ "PrematureEofInPackedFieldValue" + type_name); |
+ |
+ // EOF in the middle of packed region. |
+ ExpectParseFailureForProto( |
+ cat( tag(rep_fieldnum, WireFormatLite::WIRETYPE_LENGTH_DELIMITED), |
+ varint(1) ), |
+ "PrematureEofInPackedField" + type_name); |
+ } |
+} |
+ |
+void ConformanceTestSuite::SetFailureList(const vector<string>& failure_list) { |
+ expected_to_fail_.clear(); |
+ std::copy(failure_list.begin(), failure_list.end(), |
+ std::inserter(expected_to_fail_, expected_to_fail_.end())); |
+} |
+ |
+bool ConformanceTestSuite::CheckSetEmpty(const set<string>& set_to_check, |
+ const char* msg) { |
+ if (set_to_check.empty()) { |
+ return true; |
+ } else { |
+ StringAppendF(&output_, "\n"); |
+ StringAppendF(&output_, "%s:\n", msg); |
+ for (set<string>::const_iterator iter = set_to_check.begin(); |
+ iter != set_to_check.end(); ++iter) { |
+ StringAppendF(&output_, " %s\n", iter->c_str()); |
+ } |
+ StringAppendF(&output_, "\n"); |
+ return false; |
+ } |
+} |
+ |
+bool ConformanceTestSuite::RunSuite(ConformanceTestRunner* runner, |
+ std::string* output) { |
+ runner_ = runner; |
+ successes_ = 0; |
+ expected_failures_ = 0; |
+ skipped_.clear(); |
+ test_names_.clear(); |
+ unexpected_failing_tests_.clear(); |
+ unexpected_succeeding_tests_.clear(); |
+ type_resolver_.reset(NewTypeResolverForDescriptorPool( |
+ kTypeUrlPrefix, DescriptorPool::generated_pool())); |
+ type_url_ = GetTypeUrl(TestAllTypes::descriptor()); |
+ |
+ output_ = "\nCONFORMANCE TEST BEGIN ====================================\n\n"; |
+ |
+ for (int i = 1; i <= FieldDescriptor::MAX_TYPE; i++) { |
+ if (i == FieldDescriptor::TYPE_GROUP) continue; |
+ TestPrematureEOFForType(static_cast<FieldDescriptor::Type>(i)); |
+ } |
+ |
+ RunValidJsonTest("HelloWorld", "{\"optionalString\":\"Hello, World!\"}", |
+ "optional_string: 'Hello, World!'"); |
+ |
+ // Test field name conventions. |
+ RunValidJsonTest( |
+ "FieldNameInSnakeCase", |
+ R"({ |
+ "fieldname1": 1, |
+ "fieldName2": 2, |
+ "FieldName3": 3 |
+ })", |
+ R"( |
+ fieldname1: 1 |
+ field_name2: 2 |
+ _field_name3: 3 |
+ )"); |
+ RunValidJsonTest( |
+ "FieldNameWithNumbers", |
+ R"({ |
+ "field0name5": 5, |
+ "field0Name6": 6 |
+ })", |
+ R"( |
+ field0name5: 5 |
+ field_0_name6: 6 |
+ )"); |
+ RunValidJsonTest( |
+ "FieldNameWithMixedCases", |
+ R"({ |
+ "fieldName7": 7, |
+ "fieldName8": 8, |
+ "fieldName9": 9, |
+ "fieldName10": 10, |
+ "fIELDNAME11": 11, |
+ "fIELDName12": 12 |
+ })", |
+ R"( |
+ fieldName7: 7 |
+ FieldName8: 8 |
+ field_Name9: 9 |
+ Field_Name10: 10 |
+ FIELD_NAME11: 11 |
+ FIELD_name12: 12 |
+ )"); |
+ // Using the original proto field name in JSON is also allowed. |
+ RunValidJsonTest( |
+ "OriginalProtoFieldName", |
+ R"({ |
+ "fieldname1": 1, |
+ "field_name2": 2, |
+ "_field_name3": 3, |
+ "field0name5": 5, |
+ "field_0_name6": 6, |
+ "fieldName7": 7, |
+ "FieldName8": 8, |
+ "field_Name9": 9, |
+ "Field_Name10": 10, |
+ "FIELD_NAME11": 11, |
+ "FIELD_name12": 12 |
+ })", |
+ R"( |
+ fieldname1: 1 |
+ field_name2: 2 |
+ _field_name3: 3 |
+ field0name5: 5 |
+ field_0_name6: 6 |
+ fieldName7: 7 |
+ FieldName8: 8 |
+ field_Name9: 9 |
+ Field_Name10: 10 |
+ FIELD_NAME11: 11 |
+ FIELD_name12: 12 |
+ )"); |
+ // Field names can be escaped. |
+ RunValidJsonTest( |
+ "FieldNameEscaped", |
+ R"({"fieldn\u0061me1": 1})", |
+ "fieldname1: 1"); |
+ // Field names must be quoted (or it's not valid JSON). |
+ ExpectParseFailureForJson( |
+ "FieldNameNotQuoted", |
+ "{fieldname1: 1}"); |
+ // Trailing comma is not allowed (not valid JSON). |
+ ExpectParseFailureForJson( |
+ "TrailingCommaInAnObject", |
+ R"({"fieldname1":1,})"); |
+ // JSON doesn't support comments. |
+ ExpectParseFailureForJson( |
+ "JsonWithComments", |
+ R"({ |
+ // This is a comment. |
+ "fieldname1": 1 |
+ })"); |
+ // Duplicated field names are not allowed. |
+ ExpectParseFailureForJson( |
+ "FieldNameDuplicate", |
+ R"({ |
+ "optionalNestedMessage": {a: 1}, |
+ "optionalNestedMessage": {} |
+ })"); |
+ ExpectParseFailureForJson( |
+ "FieldNameDuplicateDifferentCasing1", |
+ R"({ |
+ "optional_nested_message": {a: 1}, |
+ "optionalNestedMessage": {} |
+ })"); |
+ ExpectParseFailureForJson( |
+ "FieldNameDuplicateDifferentCasing2", |
+ R"({ |
+ "optionalNestedMessage": {a: 1}, |
+ "optional_nested_message": {} |
+ })"); |
+ // Serializers should use lowerCamelCase by default. |
+ RunValidJsonTestWithValidator( |
+ "FieldNameInLowerCamelCase", |
+ R"({ |
+ "fieldname1": 1, |
+ "fieldName2": 2, |
+ "FieldName3": 3 |
+ })", |
+ [](const Json::Value& value) { |
+ return value.isMember("fieldname1") && |
+ value.isMember("fieldName2") && |
+ value.isMember("FieldName3"); |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "FieldNameWithNumbers", |
+ R"({ |
+ "field0name5": 5, |
+ "field0Name6": 6 |
+ })", |
+ [](const Json::Value& value) { |
+ return value.isMember("field0name5") && |
+ value.isMember("field0Name6"); |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "FieldNameWithMixedCases", |
+ R"({ |
+ "fieldName7": 7, |
+ "fieldName8": 8, |
+ "fieldName9": 9, |
+ "fieldName10": 10, |
+ "fIELDNAME11": 11, |
+ "fIELDName12": 12 |
+ })", |
+ [](const Json::Value& value) { |
+ return value.isMember("fieldName7") && |
+ value.isMember("fieldName8") && |
+ value.isMember("fieldName9") && |
+ value.isMember("fieldName10") && |
+ value.isMember("fIELDNAME11") && |
+ value.isMember("fIELDName12"); |
+ }); |
+ |
+ // Integer fields. |
+ RunValidJsonTest( |
+ "Int32FieldMaxValue", |
+ R"({"optionalInt32": 2147483647})", |
+ "optional_int32: 2147483647"); |
+ RunValidJsonTest( |
+ "Int32FieldMinValue", |
+ R"({"optionalInt32": -2147483648})", |
+ "optional_int32: -2147483648"); |
+ RunValidJsonTest( |
+ "Uint32FieldMaxValue", |
+ R"({"optionalUint32": 4294967295})", |
+ "optional_uint32: 4294967295"); |
+ RunValidJsonTest( |
+ "Int64FieldMaxValue", |
+ R"({"optionalInt64": "9223372036854775807"})", |
+ "optional_int64: 9223372036854775807"); |
+ RunValidJsonTest( |
+ "Int64FieldMinValue", |
+ R"({"optionalInt64": "-9223372036854775808"})", |
+ "optional_int64: -9223372036854775808"); |
+ RunValidJsonTest( |
+ "Uint64FieldMaxValue", |
+ R"({"optionalUint64": "18446744073709551615"})", |
+ "optional_uint64: 18446744073709551615"); |
+ RunValidJsonTest( |
+ "Int64FieldMaxValueNotQuoted", |
+ R"({"optionalInt64": 9223372036854775807})", |
+ "optional_int64: 9223372036854775807"); |
+ RunValidJsonTest( |
+ "Int64FieldMinValueNotQuoted", |
+ R"({"optionalInt64": -9223372036854775808})", |
+ "optional_int64: -9223372036854775808"); |
+ RunValidJsonTest( |
+ "Uint64FieldMaxValueNotQuoted", |
+ R"({"optionalUint64": 18446744073709551615})", |
+ "optional_uint64: 18446744073709551615"); |
+ // Values can be represented as JSON strings. |
+ RunValidJsonTest( |
+ "Int32FieldStringValue", |
+ R"({"optionalInt32": "2147483647"})", |
+ "optional_int32: 2147483647"); |
+ RunValidJsonTest( |
+ "Int32FieldStringValueEscaped", |
+ R"({"optionalInt32": "2\u003147483647"})", |
+ "optional_int32: 2147483647"); |
+ |
+ // Parsers reject out-of-bound integer values. |
+ ExpectParseFailureForJson( |
+ "Int32FieldTooLarge", |
+ R"({"optionalInt32": 2147483648})"); |
+ ExpectParseFailureForJson( |
+ "Int32FieldTooSmall", |
+ R"({"optionalInt32": -2147483649})"); |
+ ExpectParseFailureForJson( |
+ "Uint32FieldTooLarge", |
+ R"({"optionalUint32": 4294967296})"); |
+ ExpectParseFailureForJson( |
+ "Int64FieldTooLarge", |
+ R"({"optionalInt64": "9223372036854775808"})"); |
+ ExpectParseFailureForJson( |
+ "Int64FieldTooSmall", |
+ R"({"optionalInt64": "-9223372036854775809"})"); |
+ ExpectParseFailureForJson( |
+ "Uint64FieldTooLarge", |
+ R"({"optionalUint64": "18446744073709551616"})"); |
+ // Parser reject non-integer numeric values as well. |
+ ExpectParseFailureForJson( |
+ "Int32FieldNotInteger", |
+ R"({"optionalInt32": 0.5})"); |
+ ExpectParseFailureForJson( |
+ "Uint32FieldNotInteger", |
+ R"({"optionalUint32": 0.5})"); |
+ ExpectParseFailureForJson( |
+ "Int64FieldNotInteger", |
+ R"({"optionalInt64": "0.5"})"); |
+ ExpectParseFailureForJson( |
+ "Uint64FieldNotInteger", |
+ R"({"optionalUint64": "0.5"})"); |
+ |
+ // Integers but represented as float values are accepted. |
+ RunValidJsonTest( |
+ "Int32FieldFloatTrailingZero", |
+ R"({"optionalInt32": 100000.000})", |
+ "optional_int32: 100000"); |
+ RunValidJsonTest( |
+ "Int32FieldExponentialFormat", |
+ R"({"optionalInt32": 1e5})", |
+ "optional_int32: 100000"); |
+ RunValidJsonTest( |
+ "Int32FieldMaxFloatValue", |
+ R"({"optionalInt32": 2.147483647e9})", |
+ "optional_int32: 2147483647"); |
+ RunValidJsonTest( |
+ "Int32FieldMinFloatValue", |
+ R"({"optionalInt32": -2.147483648e9})", |
+ "optional_int32: -2147483648"); |
+ RunValidJsonTest( |
+ "Uint32FieldMaxFloatValue", |
+ R"({"optionalUint32": 4.294967295e9})", |
+ "optional_uint32: 4294967295"); |
+ |
+ // Parser reject non-numeric values. |
+ ExpectParseFailureForJson( |
+ "Int32FieldNotNumber", |
+ R"({"optionalInt32": "3x3"})"); |
+ ExpectParseFailureForJson( |
+ "Uint32FieldNotNumber", |
+ R"({"optionalUint32": "3x3"})"); |
+ ExpectParseFailureForJson( |
+ "Int64FieldNotNumber", |
+ R"({"optionalInt64": "3x3"})"); |
+ ExpectParseFailureForJson( |
+ "Uint64FieldNotNumber", |
+ R"({"optionalUint64": "3x3"})"); |
+ // JSON does not allow "+" on numric values. |
+ ExpectParseFailureForJson( |
+ "Int32FieldPlusSign", |
+ R"({"optionalInt32": +1})"); |
+ // JSON doesn't allow leading 0s. |
+ ExpectParseFailureForJson( |
+ "Int32FieldLeadingZero", |
+ R"({"optionalInt32": 01})"); |
+ ExpectParseFailureForJson( |
+ "Int32FieldNegativeWithLeadingZero", |
+ R"({"optionalInt32": -01})"); |
+ // String values must follow the same syntax rule. Specifically leading |
+ // or traling spaces are not allowed. |
+ ExpectParseFailureForJson( |
+ "Int32FieldLeadingSpace", |
+ R"({"optionalInt32": " 1"})"); |
+ ExpectParseFailureForJson( |
+ "Int32FieldTrailingSpace", |
+ R"({"optionalInt32": "1 "})"); |
+ |
+ // 64-bit values are serialized as strings. |
+ RunValidJsonTestWithValidator( |
+ "Int64FieldBeString", |
+ R"({"optionalInt64": 1})", |
+ [](const Json::Value& value) { |
+ return value["optionalInt64"].type() == Json::stringValue && |
+ value["optionalInt64"].asString() == "1"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "Uint64FieldBeString", |
+ R"({"optionalUint64": 1})", |
+ [](const Json::Value& value) { |
+ return value["optionalUint64"].type() == Json::stringValue && |
+ value["optionalUint64"].asString() == "1"; |
+ }); |
+ |
+ // Bool fields. |
+ RunValidJsonTest( |
+ "BoolFieldTrue", |
+ R"({"optionalBool":true})", |
+ "optional_bool: true"); |
+ RunValidJsonTest( |
+ "BoolFieldFalse", |
+ R"({"optionalBool":false})", |
+ "optional_bool: false"); |
+ |
+ // Other forms are not allowed. |
+ ExpectParseFailureForJson( |
+ "BoolFieldIntegerZero", |
+ R"({"optionalBool":0})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldIntegerOne", |
+ R"({"optionalBool":1})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldCamelCaseTrue", |
+ R"({"optionalBool":True})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldCamelCaseFalse", |
+ R"({"optionalBool":False})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldAllCapitalTrue", |
+ R"({"optionalBool":TRUE})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldAllCapitalFalse", |
+ R"({"optionalBool":FALSE})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldDoubleQuotedTrue", |
+ R"({"optionalBool":"true"})"); |
+ ExpectParseFailureForJson( |
+ "BoolFieldDoubleQuotedFalse", |
+ R"({"optionalBool":"false"})"); |
+ |
+ // Float fields. |
+ RunValidJsonTest( |
+ "FloatFieldMinPositiveValue", |
+ R"({"optionalFloat": 1.175494e-38})", |
+ "optional_float: 1.175494e-38"); |
+ RunValidJsonTest( |
+ "FloatFieldMaxNegativeValue", |
+ R"({"optionalFloat": -1.175494e-38})", |
+ "optional_float: -1.175494e-38"); |
+ RunValidJsonTest( |
+ "FloatFieldMaxPositiveValue", |
+ R"({"optionalFloat": 3.402823e+38})", |
+ "optional_float: 3.402823e+38"); |
+ RunValidJsonTest( |
+ "FloatFieldMinNegativeValue", |
+ R"({"optionalFloat": 3.402823e+38})", |
+ "optional_float: 3.402823e+38"); |
+ // Values can be quoted. |
+ RunValidJsonTest( |
+ "FloatFieldQuotedValue", |
+ R"({"optionalFloat": "1"})", |
+ "optional_float: 1"); |
+ // Special values. |
+ RunValidJsonTest( |
+ "FloatFieldNan", |
+ R"({"optionalFloat": "NaN"})", |
+ "optional_float: nan"); |
+ RunValidJsonTest( |
+ "FloatFieldInfinity", |
+ R"({"optionalFloat": "Infinity"})", |
+ "optional_float: inf"); |
+ RunValidJsonTest( |
+ "FloatFieldNegativeInfinity", |
+ R"({"optionalFloat": "-Infinity"})", |
+ "optional_float: -inf"); |
+ // Non-cannonical Nan will be correctly normalized. |
+ { |
+ TestAllTypes message; |
+ // IEEE floating-point standard 32-bit quiet NaN: |
+ // 0111 1111 1xxx xxxx xxxx xxxx xxxx xxxx |
+ message.set_optional_float( |
+ WireFormatLite::DecodeFloat(0x7FA12345)); |
+ RunValidJsonTestWithProtobufInput( |
+ "FloatFieldNormalizeQuietNan", message, |
+ "optional_float: nan"); |
+ // IEEE floating-point standard 64-bit signaling NaN: |
+ // 1111 1111 1xxx xxxx xxxx xxxx xxxx xxxx |
+ message.set_optional_float( |
+ WireFormatLite::DecodeFloat(0xFFB54321)); |
+ RunValidJsonTestWithProtobufInput( |
+ "FloatFieldNormalizeSignalingNan", message, |
+ "optional_float: nan"); |
+ } |
+ |
+ // Special values must be quoted. |
+ ExpectParseFailureForJson( |
+ "FloatFieldNanNotQuoted", |
+ R"({"optionalFloat": NaN})"); |
+ ExpectParseFailureForJson( |
+ "FloatFieldInfinityNotQuoted", |
+ R"({"optionalFloat": Infinity})"); |
+ ExpectParseFailureForJson( |
+ "FloatFieldNegativeInfinityNotQuoted", |
+ R"({"optionalFloat": -Infinity})"); |
+ // Parsers should reject out-of-bound values. |
+ ExpectParseFailureForJson( |
+ "FloatFieldTooSmall", |
+ R"({"optionalFloat": -3.502823e+38})"); |
+ ExpectParseFailureForJson( |
+ "FloatFieldTooLarge", |
+ R"({"optionalFloat": 3.502823e+38})"); |
+ |
+ // Double fields. |
+ RunValidJsonTest( |
+ "DoubleFieldMinPositiveValue", |
+ R"({"optionalDouble": 2.22507e-308})", |
+ "optional_double: 2.22507e-308"); |
+ RunValidJsonTest( |
+ "DoubleFieldMaxNegativeValue", |
+ R"({"optionalDouble": -2.22507e-308})", |
+ "optional_double: -2.22507e-308"); |
+ RunValidJsonTest( |
+ "DoubleFieldMaxPositiveValue", |
+ R"({"optionalDouble": 1.79769e+308})", |
+ "optional_double: 1.79769e+308"); |
+ RunValidJsonTest( |
+ "DoubleFieldMinNegativeValue", |
+ R"({"optionalDouble": -1.79769e+308})", |
+ "optional_double: -1.79769e+308"); |
+ // Values can be quoted. |
+ RunValidJsonTest( |
+ "DoubleFieldQuotedValue", |
+ R"({"optionalDouble": "1"})", |
+ "optional_double: 1"); |
+ // Speical values. |
+ RunValidJsonTest( |
+ "DoubleFieldNan", |
+ R"({"optionalDouble": "NaN"})", |
+ "optional_double: nan"); |
+ RunValidJsonTest( |
+ "DoubleFieldInfinity", |
+ R"({"optionalDouble": "Infinity"})", |
+ "optional_double: inf"); |
+ RunValidJsonTest( |
+ "DoubleFieldNegativeInfinity", |
+ R"({"optionalDouble": "-Infinity"})", |
+ "optional_double: -inf"); |
+ // Non-cannonical Nan will be correctly normalized. |
+ { |
+ TestAllTypes message; |
+ message.set_optional_double( |
+ WireFormatLite::DecodeDouble(0x7FFA123456789ABCLL)); |
+ RunValidJsonTestWithProtobufInput( |
+ "DoubleFieldNormalizeQuietNan", message, |
+ "optional_double: nan"); |
+ message.set_optional_double( |
+ WireFormatLite::DecodeDouble(0xFFFBCBA987654321LL)); |
+ RunValidJsonTestWithProtobufInput( |
+ "DoubleFieldNormalizeSignalingNan", message, |
+ "optional_double: nan"); |
+ } |
+ |
+ // Special values must be quoted. |
+ ExpectParseFailureForJson( |
+ "DoubleFieldNanNotQuoted", |
+ R"({"optionalDouble": NaN})"); |
+ ExpectParseFailureForJson( |
+ "DoubleFieldInfinityNotQuoted", |
+ R"({"optionalDouble": Infinity})"); |
+ ExpectParseFailureForJson( |
+ "DoubleFieldNegativeInfinityNotQuoted", |
+ R"({"optionalDouble": -Infinity})"); |
+ |
+ // Parsers should reject out-of-bound values. |
+ ExpectParseFailureForJson( |
+ "DoubleFieldTooSmall", |
+ R"({"optionalDouble": -1.89769e+308})"); |
+ ExpectParseFailureForJson( |
+ "DoubleFieldTooLarge", |
+ R"({"optionalDouble": +1.89769e+308})"); |
+ |
+ // Enum fields. |
+ RunValidJsonTest( |
+ "EnumField", |
+ R"({"optionalNestedEnum": "FOO"})", |
+ "optional_nested_enum: FOO"); |
+ // Enum values must be represented as strings. |
+ ExpectParseFailureForJson( |
+ "EnumFieldNotQuoted", |
+ R"({"optionalNestedEnum": FOO})"); |
+ // Numeric values are allowed. |
+ RunValidJsonTest( |
+ "EnumFieldNumericValueZero", |
+ R"({"optionalNestedEnum": 0})", |
+ "optional_nested_enum: FOO"); |
+ RunValidJsonTest( |
+ "EnumFieldNumericValueNonZero", |
+ R"({"optionalNestedEnum": 1})", |
+ "optional_nested_enum: BAR"); |
+ // Unknown enum values are represented as numeric values. |
+ RunValidJsonTestWithValidator( |
+ "EnumFieldUnknownValue", |
+ R"({"optionalNestedEnum": 123})", |
+ [](const Json::Value& value) { |
+ return value["optionalNestedEnum"].type() == Json::intValue && |
+ value["optionalNestedEnum"].asInt() == 123; |
+ }); |
+ |
+ // String fields. |
+ RunValidJsonTest( |
+ "StringField", |
+ R"({"optionalString": "Hello world!"})", |
+ "optional_string: \"Hello world!\""); |
+ RunValidJsonTest( |
+ "StringFieldUnicode", |
+ // Google in Chinese. |
+ R"({"optionalString": "谷歌"})", |
+ R"(optional_string: "谷歌")"); |
+ RunValidJsonTest( |
+ "StringFieldEscape", |
+ R"({"optionalString": "\"\\\/\b\f\n\r\t"})", |
+ R"(optional_string: "\"\\/\b\f\n\r\t")"); |
+ RunValidJsonTest( |
+ "StringFieldUnicodeEscape", |
+ R"({"optionalString": "\u8C37\u6B4C"})", |
+ R"(optional_string: "谷歌")"); |
+ RunValidJsonTest( |
+ "StringFieldUnicodeEscapeWithLowercaseHexLetters", |
+ R"({"optionalString": "\u8c37\u6b4c"})", |
+ R"(optional_string: "谷歌")"); |
+ RunValidJsonTest( |
+ "StringFieldSurrogatePair", |
+ // The character is an emoji: grinning face with smiling eyes. 😁 |
+ R"({"optionalString": "\uD83D\uDE01"})", |
+ R"(optional_string: "\xF0\x9F\x98\x81")"); |
+ |
+ // Unicode escapes must start with "\u" (lowercase u). |
+ ExpectParseFailureForJson( |
+ "StringFieldUppercaseEscapeLetter", |
+ R"({"optionalString": "\U8C37\U6b4C"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldInvalidEscape", |
+ R"({"optionalString": "\uXXXX\u6B4C"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldUnterminatedEscape", |
+ R"({"optionalString": "\u8C3"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldUnpairedHighSurrogate", |
+ R"({"optionalString": "\uD800"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldUnpairedLowSurrogate", |
+ R"({"optionalString": "\uDC00"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldSurrogateInWrongOrder", |
+ R"({"optionalString": "\uDE01\uD83D"})"); |
+ ExpectParseFailureForJson( |
+ "StringFieldNotAString", |
+ R"({"optionalString": 12345})"); |
+ |
+ // Bytes fields. |
+ RunValidJsonTest( |
+ "BytesField", |
+ R"({"optionalBytes": "AQI="})", |
+ R"(optional_bytes: "\x01\x02")"); |
+ ExpectParseFailureForJson( |
+ "BytesFieldNoPadding", |
+ R"({"optionalBytes": "AQI"})"); |
+ ExpectParseFailureForJson( |
+ "BytesFieldInvalidBase64Characters", |
+ R"({"optionalBytes": "-_=="})"); |
+ |
+ // Message fields. |
+ RunValidJsonTest( |
+ "MessageField", |
+ R"({"optionalNestedMessage": {"a": 1234}})", |
+ "optional_nested_message: {a: 1234}"); |
+ |
+ // Oneof fields. |
+ ExpectParseFailureForJson( |
+ "OneofFieldDuplicate", |
+ R"({"oneofUint32": 1, "oneofString": "test"})"); |
+ |
+ // Repeated fields. |
+ RunValidJsonTest( |
+ "PrimitiveRepeatedField", |
+ R"({"repeatedInt32": [1, 2, 3, 4]})", |
+ "repeated_int32: [1, 2, 3, 4]"); |
+ RunValidJsonTest( |
+ "EnumRepeatedField", |
+ R"({"repeatedNestedEnum": ["FOO", "BAR", "BAZ"]})", |
+ "repeated_nested_enum: [FOO, BAR, BAZ]"); |
+ RunValidJsonTest( |
+ "StringRepeatedField", |
+ R"({"repeatedString": ["Hello", "world"]})", |
+ R"(repeated_string: ["Hello", "world"])"); |
+ RunValidJsonTest( |
+ "BytesRepeatedField", |
+ R"({"repeatedBytes": ["AAEC", "AQI="]})", |
+ R"(repeated_bytes: ["\x00\x01\x02", "\x01\x02"])"); |
+ RunValidJsonTest( |
+ "MessageRepeatedField", |
+ R"({"repeatedNestedMessage": [{"a": 1234}, {"a": 5678}]})", |
+ "repeated_nested_message: {a: 1234}" |
+ "repeated_nested_message: {a: 5678}"); |
+ |
+ // Repeated field elements are of incorrect type. |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingIntegersGotBool", |
+ R"({"repeatedInt32": [1, false, 3, 4]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingIntegersGotString", |
+ R"({"repeatedInt32": [1, 2, "name", 4]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingIntegersGotMessage", |
+ R"({"repeatedInt32": [1, 2, 3, {"a": 4}]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingStringsGotInt", |
+ R"({"repeatedString": ["1", 2, "3", "4"]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingStringsGotBool", |
+ R"({"repeatedString": ["1", "2", false, "4"]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingStringsGotMessage", |
+ R"({"repeatedString": ["1", 2, "3", {"a": 4}]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingMessagesGotInt", |
+ R"({"repeatedNestedMessage": [{"a": 1}, 2]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingMessagesGotBool", |
+ R"({"repeatedNestedMessage": [{"a": 1}, false]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldWrongElementTypeExpectingMessagesGotString", |
+ R"({"repeatedNestedMessage": [{"a": 1}, "2"]})"); |
+ // Trailing comma in the repeated field is not allowed. |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldTrailingComma", |
+ R"({"repeatedInt32": [1, 2, 3, 4,]})"); |
+ |
+ // Map fields. |
+ RunValidJsonTest( |
+ "Int32MapField", |
+ R"({"mapInt32Int32": {"1": 2, "3": 4}})", |
+ "map_int32_int32: {key: 1 value: 2}" |
+ "map_int32_int32: {key: 3 value: 4}"); |
+ ExpectParseFailureForJson( |
+ "Int32MapFieldKeyNotQuoted", |
+ R"({"mapInt32Int32": {1: 2, 3: 4}})"); |
+ RunValidJsonTest( |
+ "Uint32MapField", |
+ R"({"mapUint32Uint32": {"1": 2, "3": 4}})", |
+ "map_uint32_uint32: {key: 1 value: 2}" |
+ "map_uint32_uint32: {key: 3 value: 4}"); |
+ ExpectParseFailureForJson( |
+ "Uint32MapFieldKeyNotQuoted", |
+ R"({"mapUint32Uint32": {1: 2, 3: 4}})"); |
+ RunValidJsonTest( |
+ "Int64MapField", |
+ R"({"mapInt64Int64": {"1": 2, "3": 4}})", |
+ "map_int64_int64: {key: 1 value: 2}" |
+ "map_int64_int64: {key: 3 value: 4}"); |
+ ExpectParseFailureForJson( |
+ "Int64MapFieldKeyNotQuoted", |
+ R"({"mapInt64Int64": {1: 2, 3: 4}})"); |
+ RunValidJsonTest( |
+ "Uint64MapField", |
+ R"({"mapUint64Uint64": {"1": 2, "3": 4}})", |
+ "map_uint64_uint64: {key: 1 value: 2}" |
+ "map_uint64_uint64: {key: 3 value: 4}"); |
+ ExpectParseFailureForJson( |
+ "Uint64MapFieldKeyNotQuoted", |
+ R"({"mapUint64Uint64": {1: 2, 3: 4}})"); |
+ RunValidJsonTest( |
+ "BoolMapField", |
+ R"({"mapBoolBool": {"true": true, "false": false}})", |
+ "map_bool_bool: {key: true value: true}" |
+ "map_bool_bool: {key: false value: false}"); |
+ ExpectParseFailureForJson( |
+ "BoolMapFieldKeyNotQuoted", |
+ R"({"mapBoolBool": {true: true, false: false}})"); |
+ RunValidJsonTest( |
+ "MessageMapField", |
+ R"({ |
+ "mapStringNestedMessage": { |
+ "hello": {"a": 1234}, |
+ "world": {"a": 5678} |
+ } |
+ })", |
+ R"( |
+ map_string_nested_message: { |
+ key: "hello" |
+ value: {a: 1234} |
+ } |
+ map_string_nested_message: { |
+ key: "world" |
+ value: {a: 5678} |
+ } |
+ )"); |
+ // Since Map keys are represented as JSON strings, escaping should be allowed. |
+ RunValidJsonTest( |
+ "Int32MapEscapedKey", |
+ R"({"mapInt32Int32": {"\u0031": 2}})", |
+ "map_int32_int32: {key: 1 value: 2}"); |
+ RunValidJsonTest( |
+ "Int64MapEscapedKey", |
+ R"({"mapInt64Int64": {"\u0031": 2}})", |
+ "map_int64_int64: {key: 1 value: 2}"); |
+ RunValidJsonTest( |
+ "BoolMapEscapedKey", |
+ R"({"mapBoolBool": {"tr\u0075e": true}})", |
+ "map_bool_bool: {key: true value: true}"); |
+ |
+ // "null" is accepted for all fields types. |
+ RunValidJsonTest( |
+ "AllFieldAcceptNull", |
+ R"({ |
+ "optionalInt32": null, |
+ "optionalInt64": null, |
+ "optionalUint32": null, |
+ "optionalUint64": null, |
+ "optionalBool": null, |
+ "optionalString": null, |
+ "optionalBytes": null, |
+ "optionalNestedEnum": null, |
+ "optionalNestedMessage": null, |
+ "repeatedInt32": null, |
+ "repeatedInt64": null, |
+ "repeatedUint32": null, |
+ "repeatedUint64": null, |
+ "repeatedBool": null, |
+ "repeatedString": null, |
+ "repeatedBytes": null, |
+ "repeatedNestedEnum": null, |
+ "repeatedNestedMessage": null, |
+ "mapInt32Int32": null, |
+ "mapBoolBool": null, |
+ "mapStringNestedMessage": null |
+ })", |
+ ""); |
+ |
+ // Repeated field elements cannot be null. |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldPrimitiveElementIsNull", |
+ R"({"repeatedInt32": [1, null, 2]})"); |
+ ExpectParseFailureForJson( |
+ "RepeatedFieldMessageElementIsNull", |
+ R"({"repeatedNestedMessage": [{"a":1}, null, {"a":2}]})"); |
+ // Map field keys cannot be null. |
+ ExpectParseFailureForJson( |
+ "MapFieldKeyIsNull", |
+ R"({"mapInt32Int32": {null: 1}})"); |
+ // Map field values cannot be null. |
+ ExpectParseFailureForJson( |
+ "MapFieldValueIsNull", |
+ R"({"mapInt32Int32": {"0": null}})"); |
+ |
+ // Wrapper types. |
+ RunValidJsonTest( |
+ "OptionalBoolWrapper", |
+ R"({"optionalBoolWrapper": false})", |
+ "optional_bool_wrapper: {value: false}"); |
+ RunValidJsonTest( |
+ "OptionalInt32Wrapper", |
+ R"({"optionalInt32Wrapper": 0})", |
+ "optional_int32_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalUint32Wrapper", |
+ R"({"optionalUint32Wrapper": 0})", |
+ "optional_uint32_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalInt64Wrapper", |
+ R"({"optionalInt64Wrapper": 0})", |
+ "optional_int64_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalUint64Wrapper", |
+ R"({"optionalUint64Wrapper": 0})", |
+ "optional_uint64_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalFloatWrapper", |
+ R"({"optionalFloatWrapper": 0})", |
+ "optional_float_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalDoubleWrapper", |
+ R"({"optionalDoubleWrapper": 0})", |
+ "optional_double_wrapper: {value: 0}"); |
+ RunValidJsonTest( |
+ "OptionalStringWrapper", |
+ R"({"optionalStringWrapper": ""})", |
+ R"(optional_string_wrapper: {value: ""})"); |
+ RunValidJsonTest( |
+ "OptionalBytesWrapper", |
+ R"({"optionalBytesWrapper": ""})", |
+ R"(optional_bytes_wrapper: {value: ""})"); |
+ RunValidJsonTest( |
+ "OptionalWrapperTypesWithNonDefaultValue", |
+ R"({ |
+ "optionalBoolWrapper": true, |
+ "optionalInt32Wrapper": 1, |
+ "optionalUint32Wrapper": 1, |
+ "optionalInt64Wrapper": "1", |
+ "optionalUint64Wrapper": "1", |
+ "optionalFloatWrapper": 1, |
+ "optionalDoubleWrapper": 1, |
+ "optionalStringWrapper": "1", |
+ "optionalBytesWrapper": "AQI=" |
+ })", |
+ R"( |
+ optional_bool_wrapper: {value: true} |
+ optional_int32_wrapper: {value: 1} |
+ optional_uint32_wrapper: {value: 1} |
+ optional_int64_wrapper: {value: 1} |
+ optional_uint64_wrapper: {value: 1} |
+ optional_float_wrapper: {value: 1} |
+ optional_double_wrapper: {value: 1} |
+ optional_string_wrapper: {value: "1"} |
+ optional_bytes_wrapper: {value: "\x01\x02"} |
+ )"); |
+ RunValidJsonTest( |
+ "RepeatedBoolWrapper", |
+ R"({"repeatedBoolWrapper": [true, false]})", |
+ "repeated_bool_wrapper: {value: true}" |
+ "repeated_bool_wrapper: {value: false}"); |
+ RunValidJsonTest( |
+ "RepeatedInt32Wrapper", |
+ R"({"repeatedInt32Wrapper": [0, 1]})", |
+ "repeated_int32_wrapper: {value: 0}" |
+ "repeated_int32_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedUint32Wrapper", |
+ R"({"repeatedUint32Wrapper": [0, 1]})", |
+ "repeated_uint32_wrapper: {value: 0}" |
+ "repeated_uint32_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedInt64Wrapper", |
+ R"({"repeatedInt64Wrapper": [0, 1]})", |
+ "repeated_int64_wrapper: {value: 0}" |
+ "repeated_int64_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedUint64Wrapper", |
+ R"({"repeatedUint64Wrapper": [0, 1]})", |
+ "repeated_uint64_wrapper: {value: 0}" |
+ "repeated_uint64_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedFloatWrapper", |
+ R"({"repeatedFloatWrapper": [0, 1]})", |
+ "repeated_float_wrapper: {value: 0}" |
+ "repeated_float_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedDoubleWrapper", |
+ R"({"repeatedDoubleWrapper": [0, 1]})", |
+ "repeated_double_wrapper: {value: 0}" |
+ "repeated_double_wrapper: {value: 1}"); |
+ RunValidJsonTest( |
+ "RepeatedStringWrapper", |
+ R"({"repeatedStringWrapper": ["", "AQI="]})", |
+ R"( |
+ repeated_string_wrapper: {value: ""} |
+ repeated_string_wrapper: {value: "AQI="} |
+ )"); |
+ RunValidJsonTest( |
+ "RepeatedBytesWrapper", |
+ R"({"repeatedBytesWrapper": ["", "AQI="]})", |
+ R"( |
+ repeated_bytes_wrapper: {value: ""} |
+ repeated_bytes_wrapper: {value: "\x01\x02"} |
+ )"); |
+ RunValidJsonTest( |
+ "WrapperTypesWithNullValue", |
+ R"({ |
+ "optionalBoolWrapper": null, |
+ "optionalInt32Wrapper": null, |
+ "optionalUint32Wrapper": null, |
+ "optionalInt64Wrapper": null, |
+ "optionalUint64Wrapper": null, |
+ "optionalFloatWrapper": null, |
+ "optionalDoubleWrapper": null, |
+ "optionalStringWrapper": null, |
+ "optionalBytesWrapper": null, |
+ "repeatedBoolWrapper": null, |
+ "repeatedInt32Wrapper": null, |
+ "repeatedUint32Wrapper": null, |
+ "repeatedInt64Wrapper": null, |
+ "repeatedUint64Wrapper": null, |
+ "repeatedFloatWrapper": null, |
+ "repeatedDoubleWrapper": null, |
+ "repeatedStringWrapper": null, |
+ "repeatedBytesWrapper": null |
+ })", |
+ ""); |
+ |
+ // Duration |
+ RunValidJsonTest( |
+ "DurationMinValue", |
+ R"({"optionalDuration": "-315576000000.999999999s"})", |
+ "optional_duration: {seconds: -315576000000 nanos: -999999999}"); |
+ RunValidJsonTest( |
+ "DurationMaxValue", |
+ R"({"optionalDuration": "315576000000.999999999s"})", |
+ "optional_duration: {seconds: 315576000000 nanos: 999999999}"); |
+ RunValidJsonTest( |
+ "DurationRepeatedValue", |
+ R"({"repeatedDuration": ["1.5s", "-1.5s"]})", |
+ "repeated_duration: {seconds: 1 nanos: 500000000}" |
+ "repeated_duration: {seconds: -1 nanos: -500000000}"); |
+ |
+ ExpectParseFailureForJson( |
+ "DurationMissingS", |
+ R"({"optionalDuration": "1"})"); |
+ ExpectParseFailureForJson( |
+ "DurationJsonInputTooSmall", |
+ R"({"optionalDuration": "-315576000001.000000000s"})"); |
+ ExpectParseFailureForJson( |
+ "DurationJsonInputTooLarge", |
+ R"({"optionalDuration": "315576000001.000000000s"})"); |
+ ExpectSerializeFailureForJson( |
+ "DurationProtoInputTooSmall", |
+ "optional_duration: {seconds: -315576000001 nanos: 0}"); |
+ ExpectSerializeFailureForJson( |
+ "DurationProtoInputTooLarge", |
+ "optional_duration: {seconds: 315576000001 nanos: 0}"); |
+ |
+ RunValidJsonTestWithValidator( |
+ "DurationHasZeroFractionalDigit", |
+ R"({"optionalDuration": "1.000000000s"})", |
+ [](const Json::Value& value) { |
+ return value["optionalDuration"].asString() == "1s"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "DurationHas3FractionalDigits", |
+ R"({"optionalDuration": "1.010000000s"})", |
+ [](const Json::Value& value) { |
+ return value["optionalDuration"].asString() == "1.010s"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "DurationHas6FractionalDigits", |
+ R"({"optionalDuration": "1.000010000s"})", |
+ [](const Json::Value& value) { |
+ return value["optionalDuration"].asString() == "1.000010s"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "DurationHas9FractionalDigits", |
+ R"({"optionalDuration": "1.000000010s"})", |
+ [](const Json::Value& value) { |
+ return value["optionalDuration"].asString() == "1.000000010s"; |
+ }); |
+ |
+ // Timestamp |
+ RunValidJsonTest( |
+ "TimestampMinValue", |
+ R"({"optionalTimestamp": "0001-01-01T00:00:00Z"})", |
+ "optional_timestamp: {seconds: -62135596800}"); |
+ RunValidJsonTest( |
+ "TimestampMaxValue", |
+ R"({"optionalTimestamp": "9999-12-31T23:59:59.999999999Z"})", |
+ "optional_timestamp: {seconds: 253402300799 nanos: 999999999}"); |
+ RunValidJsonTest( |
+ "TimestampRepeatedValue", |
+ R"({ |
+ "repeatedTimestamp": [ |
+ "0001-01-01T00:00:00Z", |
+ "9999-12-31T23:59:59.999999999Z" |
+ ] |
+ })", |
+ "repeated_timestamp: {seconds: -62135596800}" |
+ "repeated_timestamp: {seconds: 253402300799 nanos: 999999999}"); |
+ RunValidJsonTest( |
+ "TimestampWithPositiveOffset", |
+ R"({"optionalTimestamp": "1970-01-01T08:00:00+08:00"})", |
+ "optional_timestamp: {seconds: 0}"); |
+ RunValidJsonTest( |
+ "TimestampWithNegativeOffset", |
+ R"({"optionalTimestamp": "1969-12-31T16:00:00-08:00"})", |
+ "optional_timestamp: {seconds: 0}"); |
+ |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputTooSmall", |
+ R"({"optionalTimestamp": "0000-01-01T00:00:00Z"})"); |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputTooLarge", |
+ R"({"optionalTimestamp": "10000-01-01T00:00:00Z"})"); |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputMissingZ", |
+ R"({"optionalTimestamp": "0001-01-01T00:00:00"})"); |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputMissingT", |
+ R"({"optionalTimestamp": "0001-01-01 00:00:00Z"})"); |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputLowercaseZ", |
+ R"({"optionalTimestamp": "0001-01-01T00:00:00z"})"); |
+ ExpectParseFailureForJson( |
+ "TimestampJsonInputLowercaseT", |
+ R"({"optionalTimestamp": "0001-01-01t00:00:00Z"})"); |
+ ExpectSerializeFailureForJson( |
+ "TimestampProtoInputTooSmall", |
+ "optional_timestamp: {seconds: -62135596801}"); |
+ ExpectSerializeFailureForJson( |
+ "TimestampProtoInputTooLarge", |
+ "optional_timestamp: {seconds: 253402300800}"); |
+ RunValidJsonTestWithValidator( |
+ "TimestampZeroNormalized", |
+ R"({"optionalTimestamp": "1969-12-31T16:00:00-08:00"})", |
+ [](const Json::Value& value) { |
+ return value["optionalTimestamp"].asString() == |
+ "1970-01-01T00:00:00Z"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "TimestampHasZeroFractionalDigit", |
+ R"({"optionalTimestamp": "1970-01-01T00:00:00.000000000Z"})", |
+ [](const Json::Value& value) { |
+ return value["optionalTimestamp"].asString() == |
+ "1970-01-01T00:00:00Z"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "TimestampHas3FractionalDigits", |
+ R"({"optionalTimestamp": "1970-01-01T00:00:00.010000000Z"})", |
+ [](const Json::Value& value) { |
+ return value["optionalTimestamp"].asString() == |
+ "1970-01-01T00:00:00.010Z"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "TimestampHas6FractionalDigits", |
+ R"({"optionalTimestamp": "1970-01-01T00:00:00.000010000Z"})", |
+ [](const Json::Value& value) { |
+ return value["optionalTimestamp"].asString() == |
+ "1970-01-01T00:00:00.000010Z"; |
+ }); |
+ RunValidJsonTestWithValidator( |
+ "TimestampHas9FractionalDigits", |
+ R"({"optionalTimestamp": "1970-01-01T00:00:00.000000010Z"})", |
+ [](const Json::Value& value) { |
+ return value["optionalTimestamp"].asString() == |
+ "1970-01-01T00:00:00.000000010Z"; |
+ }); |
+ |
+ // FieldMask |
+ RunValidJsonTest( |
+ "FieldMask", |
+ R"({"optionalFieldMask": "foo,barBaz"})", |
+ R"(optional_field_mask: {paths: "foo" paths: "bar_baz"})"); |
+ ExpectParseFailureForJson( |
+ "FieldMaskInvalidCharacter", |
+ R"({"optionalFieldMask": "foo,bar_bar"})"); |
+ ExpectSerializeFailureForJson( |
+ "FieldMaskPathsDontRoundTrip", |
+ R"(optional_field_mask: {paths: "fooBar"})"); |
+ ExpectSerializeFailureForJson( |
+ "FieldMaskNumbersDontRoundTrip", |
+ R"(optional_field_mask: {paths: "foo_3_bar"})"); |
+ ExpectSerializeFailureForJson( |
+ "FieldMaskTooManyUnderscore", |
+ R"(optional_field_mask: {paths: "foo__bar"})"); |
+ |
+ // Struct |
+ RunValidJsonTest( |
+ "Struct", |
+ R"({ |
+ "optionalStruct": { |
+ "nullValue": null, |
+ "intValue": 1234, |
+ "boolValue": true, |
+ "doubleValue": 1234.5678, |
+ "stringValue": "Hello world!", |
+ "listValue": [1234, "5678"], |
+ "objectValue": { |
+ "value": 0 |
+ } |
+ } |
+ })", |
+ R"( |
+ optional_struct: { |
+ fields: { |
+ key: "nullValue" |
+ value: {null_value: NULL_VALUE} |
+ } |
+ fields: { |
+ key: "intValue" |
+ value: {number_value: 1234} |
+ } |
+ fields: { |
+ key: "boolValue" |
+ value: {bool_value: true} |
+ } |
+ fields: { |
+ key: "doubleValue" |
+ value: {number_value: 1234.5678} |
+ } |
+ fields: { |
+ key: "stringValue" |
+ value: {string_value: "Hello world!"} |
+ } |
+ fields: { |
+ key: "listValue" |
+ value: { |
+ list_value: { |
+ values: { |
+ number_value: 1234 |
+ } |
+ values: { |
+ string_value: "5678" |
+ } |
+ } |
+ } |
+ } |
+ fields: { |
+ key: "objectValue" |
+ value: { |
+ struct_value: { |
+ fields: { |
+ key: "value" |
+ value: { |
+ number_value: 0 |
+ } |
+ } |
+ } |
+ } |
+ } |
+ } |
+ )"); |
+ // Value |
+ RunValidJsonTest( |
+ "ValueAcceptInteger", |
+ R"({"optionalValue": 1})", |
+ "optional_value: { number_value: 1}"); |
+ RunValidJsonTest( |
+ "ValueAcceptFloat", |
+ R"({"optionalValue": 1.5})", |
+ "optional_value: { number_value: 1.5}"); |
+ RunValidJsonTest( |
+ "ValueAcceptBool", |
+ R"({"optionalValue": false})", |
+ "optional_value: { bool_value: false}"); |
+ RunValidJsonTest( |
+ "ValueAcceptNull", |
+ R"({"optionalValue": null})", |
+ "optional_value: { null_value: NULL_VALUE}"); |
+ RunValidJsonTest( |
+ "ValueAcceptString", |
+ R"({"optionalValue": "hello"})", |
+ R"(optional_value: { string_value: "hello"})"); |
+ RunValidJsonTest( |
+ "ValueAcceptList", |
+ R"({"optionalValue": [0, "hello"]})", |
+ R"( |
+ optional_value: { |
+ list_value: { |
+ values: { |
+ number_value: 0 |
+ } |
+ values: { |
+ string_value: "hello" |
+ } |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "ValueAcceptObject", |
+ R"({"optionalValue": {"value": 1}})", |
+ R"( |
+ optional_value: { |
+ struct_value: { |
+ fields: { |
+ key: "value" |
+ value: { |
+ number_value: 1 |
+ } |
+ } |
+ } |
+ } |
+ )"); |
+ |
+ // Any |
+ RunValidJsonTest( |
+ "Any", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/conformance.TestAllTypes", |
+ "optionalInt32": 12345 |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/conformance.TestAllTypes] { |
+ optional_int32: 12345 |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyNested", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Any", |
+ "value": { |
+ "@type": "type.googleapis.com/conformance.TestAllTypes", |
+ "optionalInt32": 12345 |
+ } |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Any] { |
+ [type.googleapis.com/conformance.TestAllTypes] { |
+ optional_int32: 12345 |
+ } |
+ } |
+ } |
+ )"); |
+ // The special "@type" tag is not required to appear first. |
+ RunValidJsonTest( |
+ "AnyUnorderedTypeTag", |
+ R"({ |
+ "optionalAny": { |
+ "optionalInt32": 12345, |
+ "@type": "type.googleapis.com/conformance.TestAllTypes" |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/conformance.TestAllTypes] { |
+ optional_int32: 12345 |
+ } |
+ } |
+ )"); |
+ // Well-known types in Any. |
+ RunValidJsonTest( |
+ "AnyWithInt32ValueWrapper", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Int32Value", |
+ "value": 12345 |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Int32Value] { |
+ value: 12345 |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithDuration", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Duration", |
+ "value": "1.5s" |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Duration] { |
+ seconds: 1 |
+ nanos: 500000000 |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithTimestamp", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Timestamp", |
+ "value": "1970-01-01T00:00:00Z" |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Timestamp] { |
+ seconds: 0 |
+ nanos: 0 |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithFieldMask", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.FieldMask", |
+ "value": "foo,barBaz" |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.FieldMask] { |
+ paths: ["foo", "bar_baz"] |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithStruct", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Struct", |
+ "value": { |
+ "foo": 1 |
+ } |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Struct] { |
+ fields: { |
+ key: "foo" |
+ value: { |
+ number_value: 1 |
+ } |
+ } |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithValueForJsonObject", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Value", |
+ "value": { |
+ "foo": 1 |
+ } |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Value] { |
+ struct_value: { |
+ fields: { |
+ key: "foo" |
+ value: { |
+ number_value: 1 |
+ } |
+ } |
+ } |
+ } |
+ } |
+ )"); |
+ RunValidJsonTest( |
+ "AnyWithValueForInteger", |
+ R"({ |
+ "optionalAny": { |
+ "@type": "type.googleapis.com/google.protobuf.Value", |
+ "value": 1 |
+ } |
+ })", |
+ R"( |
+ optional_any: { |
+ [type.googleapis.com/google.protobuf.Value] { |
+ number_value: 1 |
+ } |
+ } |
+ )"); |
+ |
+ bool ok = true; |
+ if (!CheckSetEmpty(expected_to_fail_, |
+ "These tests were listed in the failure list, but they " |
+ "don't exist. Remove them from the failure list")) { |
+ ok = false; |
+ } |
+ if (!CheckSetEmpty(unexpected_failing_tests_, |
+ "These tests failed. If they can't be fixed right now, " |
+ "you can add them to the failure list so the overall " |
+ "suite can succeed")) { |
+ ok = false; |
+ } |
+ |
+ // Sometimes the testee may be fixed before we update the failure list (e.g., |
+ // the testee is from a different component). We warn about this case but |
+ // don't consider it an overall test failure. |
+ CheckSetEmpty(unexpected_succeeding_tests_, |
+ "These tests succeeded, even though they were listed in " |
+ "the failure list. Remove them from the failure list"); |
+ |
+ CheckSetEmpty(skipped_, |
+ "These tests were skipped (probably because support for some " |
+ "features is not implemented)"); |
+ if (verbose_) { |
+ CheckSetEmpty(skipped_, |
+ "These tests were skipped (probably because support for some " |
+ "features is not implemented)"); |
+ } |
+ |
+ StringAppendF(&output_, |
+ "CONFORMANCE SUITE %s: %d successes, %d skipped, " |
+ "%d expected failures, %d unexpected failures.\n", |
+ ok ? "PASSED" : "FAILED", successes_, skipped_.size(), |
+ expected_failures_, unexpected_failing_tests_.size()); |
+ StringAppendF(&output_, "\n"); |
+ |
+ output->assign(output_); |
+ |
+ return ok; |
+} |
+ |
+} // namespace protobuf |
+} // namespace google |