| Index: third_party/grpc/test/cpp/end2end/thread_stress_test.cc
|
| diff --git a/third_party/grpc/test/cpp/end2end/thread_stress_test.cc b/third_party/grpc/test/cpp/end2end/thread_stress_test.cc
|
| new file mode 100644
|
| index 0000000000000000000000000000000000000000..e246c0b0e2fd4ecf6077e53662528cb59631a021
|
| --- /dev/null
|
| +++ b/third_party/grpc/test/cpp/end2end/thread_stress_test.cc
|
| @@ -0,0 +1,350 @@
|
| +/*
|
| + *
|
| + * Copyright 2015-2016, Google Inc.
|
| + * All rights reserved.
|
| + *
|
| + * Redistribution and use in source and binary forms, with or without
|
| + * modification, are permitted provided that the following conditions are
|
| + * met:
|
| + *
|
| + * * Redistributions of source code must retain the above copyright
|
| + * notice, this list of conditions and the following disclaimer.
|
| + * * Redistributions in binary form must reproduce the above
|
| + * copyright notice, this list of conditions and the following disclaimer
|
| + * in the documentation and/or other materials provided with the
|
| + * distribution.
|
| + * * Neither the name of Google Inc. nor the names of its
|
| + * contributors may be used to endorse or promote products derived from
|
| + * this software without specific prior written permission.
|
| + *
|
| + * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
| + * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
| + * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
| + * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
| + * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
| + * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
| + * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
| + * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
| + * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
| + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
| + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
| + *
|
| + */
|
| +
|
| +#include <mutex>
|
| +#include <thread>
|
| +
|
| +#include <grpc++/channel.h>
|
| +#include <grpc++/client_context.h>
|
| +#include <grpc++/create_channel.h>
|
| +#include <grpc++/server.h>
|
| +#include <grpc++/server_builder.h>
|
| +#include <grpc++/server_context.h>
|
| +#include <grpc/grpc.h>
|
| +#include <grpc/support/thd.h>
|
| +#include <grpc/support/time.h>
|
| +#include <gtest/gtest.h>
|
| +
|
| +#include "src/core/surface/api_trace.h"
|
| +#include "src/proto/grpc/testing/duplicate/echo_duplicate.grpc.pb.h"
|
| +#include "src/proto/grpc/testing/echo.grpc.pb.h"
|
| +#include "test/core/util/port.h"
|
| +#include "test/core/util/test_config.h"
|
| +
|
| +using grpc::testing::EchoRequest;
|
| +using grpc::testing::EchoResponse;
|
| +using std::chrono::system_clock;
|
| +
|
| +const int kNumThreads = 100; // Number of threads
|
| +const int kNumAsyncSendThreads = 2;
|
| +const int kNumAsyncReceiveThreads = 50;
|
| +const int kNumRpcs = 1000; // Number of RPCs per thread
|
| +
|
| +namespace grpc {
|
| +namespace testing {
|
| +
|
| +namespace {
|
| +
|
| +// When echo_deadline is requested, deadline seen in the ServerContext is set in
|
| +// the response in seconds.
|
| +void MaybeEchoDeadline(ServerContext* context, const EchoRequest* request,
|
| + EchoResponse* response) {
|
| + if (request->has_param() && request->param().echo_deadline()) {
|
| + gpr_timespec deadline = gpr_inf_future(GPR_CLOCK_REALTIME);
|
| + if (context->deadline() != system_clock::time_point::max()) {
|
| + Timepoint2Timespec(context->deadline(), &deadline);
|
| + }
|
| + response->mutable_param()->set_request_deadline(deadline.tv_sec);
|
| + }
|
| +}
|
| +
|
| +} // namespace
|
| +
|
| +class TestServiceImpl : public ::grpc::testing::EchoTestService::Service {
|
| + public:
|
| + TestServiceImpl() : signal_client_(false) {}
|
| +
|
| + Status Echo(ServerContext* context, const EchoRequest* request,
|
| + EchoResponse* response) GRPC_OVERRIDE {
|
| + response->set_message(request->message());
|
| + MaybeEchoDeadline(context, request, response);
|
| + if (request->has_param() && request->param().client_cancel_after_us()) {
|
| + {
|
| + unique_lock<mutex> lock(mu_);
|
| + signal_client_ = true;
|
| + }
|
| + while (!context->IsCancelled()) {
|
| + gpr_sleep_until(gpr_time_add(
|
| + gpr_now(GPR_CLOCK_REALTIME),
|
| + gpr_time_from_micros(request->param().client_cancel_after_us(),
|
| + GPR_TIMESPAN)));
|
| + }
|
| + return Status::CANCELLED;
|
| + } else if (request->has_param() &&
|
| + request->param().server_cancel_after_us()) {
|
| + gpr_sleep_until(gpr_time_add(
|
| + gpr_now(GPR_CLOCK_REALTIME),
|
| + gpr_time_from_micros(request->param().server_cancel_after_us(),
|
| + GPR_TIMESPAN)));
|
| + return Status::CANCELLED;
|
| + } else {
|
| + EXPECT_FALSE(context->IsCancelled());
|
| + }
|
| + return Status::OK;
|
| + }
|
| +
|
| + // Unimplemented is left unimplemented to test the returned error.
|
| +
|
| + Status RequestStream(ServerContext* context,
|
| + ServerReader<EchoRequest>* reader,
|
| + EchoResponse* response) GRPC_OVERRIDE {
|
| + EchoRequest request;
|
| + response->set_message("");
|
| + while (reader->Read(&request)) {
|
| + response->mutable_message()->append(request.message());
|
| + }
|
| + return Status::OK;
|
| + }
|
| +
|
| + // Return 3 messages.
|
| + // TODO(yangg) make it generic by adding a parameter into EchoRequest
|
| + Status ResponseStream(ServerContext* context, const EchoRequest* request,
|
| + ServerWriter<EchoResponse>* writer) GRPC_OVERRIDE {
|
| + EchoResponse response;
|
| + response.set_message(request->message() + "0");
|
| + writer->Write(response);
|
| + response.set_message(request->message() + "1");
|
| + writer->Write(response);
|
| + response.set_message(request->message() + "2");
|
| + writer->Write(response);
|
| +
|
| + return Status::OK;
|
| + }
|
| +
|
| + Status BidiStream(ServerContext* context,
|
| + ServerReaderWriter<EchoResponse, EchoRequest>* stream)
|
| + GRPC_OVERRIDE {
|
| + EchoRequest request;
|
| + EchoResponse response;
|
| + while (stream->Read(&request)) {
|
| + gpr_log(GPR_INFO, "recv msg %s", request.message().c_str());
|
| + response.set_message(request.message());
|
| + stream->Write(response);
|
| + }
|
| + return Status::OK;
|
| + }
|
| +
|
| + bool signal_client() {
|
| + unique_lock<mutex> lock(mu_);
|
| + return signal_client_;
|
| + }
|
| +
|
| + private:
|
| + bool signal_client_;
|
| + mutex mu_;
|
| +};
|
| +
|
| +class TestServiceImplDupPkg
|
| + : public ::grpc::testing::duplicate::EchoTestService::Service {
|
| + public:
|
| + Status Echo(ServerContext* context, const EchoRequest* request,
|
| + EchoResponse* response) GRPC_OVERRIDE {
|
| + response->set_message("no package");
|
| + return Status::OK;
|
| + }
|
| +};
|
| +
|
| +class CommonStressTest {
|
| + public:
|
| + CommonStressTest() : kMaxMessageSize_(8192) {}
|
| + void SetUp() {
|
| + int port = grpc_pick_unused_port_or_die();
|
| + server_address_ << "localhost:" << port;
|
| + // Setup server
|
| + ServerBuilder builder;
|
| + builder.AddListeningPort(server_address_.str(),
|
| + InsecureServerCredentials());
|
| + builder.RegisterService(&service_);
|
| + builder.SetMaxMessageSize(
|
| + kMaxMessageSize_); // For testing max message size.
|
| + builder.RegisterService(&dup_pkg_service_);
|
| + server_ = builder.BuildAndStart();
|
| + }
|
| + void TearDown() { server_->Shutdown(); }
|
| + void ResetStub() {
|
| + std::shared_ptr<Channel> channel =
|
| + CreateChannel(server_address_.str(), InsecureChannelCredentials());
|
| + stub_ = grpc::testing::EchoTestService::NewStub(channel);
|
| + }
|
| + grpc::testing::EchoTestService::Stub* GetStub() { return stub_.get(); }
|
| +
|
| + private:
|
| + std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_;
|
| + std::unique_ptr<Server> server_;
|
| + std::ostringstream server_address_;
|
| + const int kMaxMessageSize_;
|
| + TestServiceImpl service_;
|
| + TestServiceImplDupPkg dup_pkg_service_;
|
| +};
|
| +
|
| +class End2endTest : public ::testing::Test {
|
| + protected:
|
| + End2endTest() {}
|
| + void SetUp() GRPC_OVERRIDE { common_.SetUp(); }
|
| + void TearDown() GRPC_OVERRIDE { common_.TearDown(); }
|
| + void ResetStub() { common_.ResetStub(); }
|
| +
|
| + CommonStressTest common_;
|
| +};
|
| +
|
| +static void SendRpc(grpc::testing::EchoTestService::Stub* stub, int num_rpcs) {
|
| + EchoRequest request;
|
| + EchoResponse response;
|
| + request.set_message("Hello");
|
| +
|
| + for (int i = 0; i < num_rpcs; ++i) {
|
| + ClientContext context;
|
| + Status s = stub->Echo(&context, request, &response);
|
| + EXPECT_EQ(response.message(), request.message());
|
| + EXPECT_TRUE(s.ok());
|
| + }
|
| +}
|
| +
|
| +TEST_F(End2endTest, ThreadStress) {
|
| + common_.ResetStub();
|
| + std::vector<std::thread*> threads;
|
| + for (int i = 0; i < kNumThreads; ++i) {
|
| + threads.push_back(new std::thread(SendRpc, common_.GetStub(), kNumRpcs));
|
| + }
|
| + for (int i = 0; i < kNumThreads; ++i) {
|
| + threads[i]->join();
|
| + delete threads[i];
|
| + }
|
| +}
|
| +
|
| +class AsyncClientEnd2endTest : public ::testing::Test {
|
| + protected:
|
| + AsyncClientEnd2endTest() : rpcs_outstanding_(0) {}
|
| +
|
| + void SetUp() GRPC_OVERRIDE { common_.SetUp(); }
|
| + void TearDown() GRPC_OVERRIDE {
|
| + void* ignored_tag;
|
| + bool ignored_ok;
|
| + while (cq_.Next(&ignored_tag, &ignored_ok))
|
| + ;
|
| + common_.TearDown();
|
| + }
|
| +
|
| + void Wait() {
|
| + unique_lock<mutex> l(mu_);
|
| + while (rpcs_outstanding_ != 0) {
|
| + cv_.wait(l);
|
| + }
|
| +
|
| + cq_.Shutdown();
|
| + }
|
| +
|
| + struct AsyncClientCall {
|
| + EchoResponse response;
|
| + ClientContext context;
|
| + Status status;
|
| + std::unique_ptr<ClientAsyncResponseReader<EchoResponse>> response_reader;
|
| + };
|
| +
|
| + void AsyncSendRpc(int num_rpcs) {
|
| + for (int i = 0; i < num_rpcs; ++i) {
|
| + AsyncClientCall* call = new AsyncClientCall;
|
| + EchoRequest request;
|
| + request.set_message("Hello: " + std::to_string(i));
|
| + call->response_reader =
|
| + common_.GetStub()->AsyncEcho(&call->context, request, &cq_);
|
| + call->response_reader->Finish(&call->response, &call->status,
|
| + (void*)call);
|
| +
|
| + unique_lock<mutex> l(mu_);
|
| + rpcs_outstanding_++;
|
| + }
|
| + }
|
| +
|
| + void AsyncCompleteRpc() {
|
| + while (true) {
|
| + void* got_tag;
|
| + bool ok = false;
|
| + if (!cq_.Next(&got_tag, &ok)) break;
|
| + AsyncClientCall* call = static_cast<AsyncClientCall*>(got_tag);
|
| + if (!ok) {
|
| + gpr_log(GPR_DEBUG, "Error: %d", call->status.error_code());
|
| + }
|
| + delete call;
|
| +
|
| + bool notify;
|
| + {
|
| + unique_lock<mutex> l(mu_);
|
| + rpcs_outstanding_--;
|
| + notify = (rpcs_outstanding_ == 0);
|
| + }
|
| + if (notify) {
|
| + cv_.notify_all();
|
| + }
|
| + }
|
| + }
|
| +
|
| + CommonStressTest common_;
|
| + CompletionQueue cq_;
|
| + mutex mu_;
|
| + condition_variable cv_;
|
| + int rpcs_outstanding_;
|
| +};
|
| +
|
| +TEST_F(AsyncClientEnd2endTest, ThreadStress) {
|
| + common_.ResetStub();
|
| + std::vector<std::thread*> send_threads, completion_threads;
|
| + for (int i = 0; i < kNumAsyncReceiveThreads; ++i) {
|
| + completion_threads.push_back(new std::thread(
|
| + &AsyncClientEnd2endTest_ThreadStress_Test::AsyncCompleteRpc, this));
|
| + }
|
| + for (int i = 0; i < kNumAsyncSendThreads; ++i) {
|
| + send_threads.push_back(
|
| + new std::thread(&AsyncClientEnd2endTest_ThreadStress_Test::AsyncSendRpc,
|
| + this, kNumRpcs));
|
| + }
|
| + for (int i = 0; i < kNumAsyncSendThreads; ++i) {
|
| + send_threads[i]->join();
|
| + delete send_threads[i];
|
| + }
|
| +
|
| + Wait();
|
| + for (int i = 0; i < kNumAsyncReceiveThreads; ++i) {
|
| + completion_threads[i]->join();
|
| + delete completion_threads[i];
|
| + }
|
| +}
|
| +
|
| +} // namespace testing
|
| +} // namespace grpc
|
| +
|
| +int main(int argc, char** argv) {
|
| + grpc_test_init(argc, argv);
|
| + ::testing::InitGoogleTest(&argc, argv);
|
| + return RUN_ALL_TESTS();
|
| +}
|
|
|