/* * * Copyright 2015 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include "src/core/lib/gpr/env.h" #include "src/core/lib/surface/api_trace.h" #include "src/proto/grpc/testing/duplicate/echo_duplicate.grpc.pb.h" #include "src/proto/grpc/testing/echo.grpc.pb.h" #include "test/core/util/port.h" #include "test/core/util/test_config.h" #include using grpc::testing::EchoRequest; using grpc::testing::EchoResponse; const int kNumThreads = 100; // Number of threads const int kNumAsyncSendThreads = 2; const int kNumAsyncReceiveThreads = 50; const int kNumAsyncServerThreads = 50; const int kNumRpcs = 1000; // Number of RPCs per thread namespace grpc { namespace testing { class TestServiceImpl : public ::grpc::testing::EchoTestService::Service { public: TestServiceImpl() {} Status Echo(ServerContext* /*context*/, const EchoRequest* request, EchoResponse* response) override { response->set_message(request->message()); return Status::OK; } }; template class CommonStressTest { public: CommonStressTest() : kMaxMessageSize_(8192) { #if TARGET_OS_IPHONE // Workaround Apple CFStream bug gpr_setenv("grpc_cfstream", "0"); #endif } virtual ~CommonStressTest() {} virtual void SetUp() = 0; virtual void TearDown() = 0; virtual void ResetStub() = 0; virtual bool AllowExhaustion() = 0; grpc::testing::EchoTestService::Stub* GetStub() { return stub_.get(); } protected: std::unique_ptr stub_; std::unique_ptr server_; virtual void SetUpStart(ServerBuilder* builder, Service* service) = 0; void SetUpStartCommon(ServerBuilder* builder, Service* service) { builder->RegisterService(service); builder->SetMaxMessageSize( kMaxMessageSize_); // For testing max message size. } void SetUpEnd(ServerBuilder* builder) { server_ = builder->BuildAndStart(); } void TearDownStart() { server_->Shutdown(); } void TearDownEnd() {} private: const int kMaxMessageSize_; }; template class CommonStressTestInsecure : public CommonStressTest { public: void ResetStub() override { std::shared_ptr channel = grpc::CreateChannel( server_address_.str(), InsecureChannelCredentials()); this->stub_ = grpc::testing::EchoTestService::NewStub(channel); } bool AllowExhaustion() override { return false; } protected: void SetUpStart(ServerBuilder* builder, Service* service) override { int port = grpc_pick_unused_port_or_die(); this->server_address_ << "localhost:" << port; // Setup server builder->AddListeningPort(server_address_.str(), InsecureServerCredentials()); this->SetUpStartCommon(builder, service); } private: std::ostringstream server_address_; }; template class CommonStressTestInproc : public CommonStressTest { public: void ResetStub() override { ChannelArguments args; std::shared_ptr channel = this->server_->InProcessChannel(args); this->stub_ = grpc::testing::EchoTestService::NewStub(channel); } bool AllowExhaustion() override { return allow_resource_exhaustion; } protected: void SetUpStart(ServerBuilder* builder, Service* service) override { this->SetUpStartCommon(builder, service); } }; template class CommonStressTestSyncServer : public BaseClass { public: void SetUp() override { ServerBuilder builder; this->SetUpStart(&builder, &service_); this->SetUpEnd(&builder); } void TearDown() override { this->TearDownStart(); this->TearDownEnd(); } private: TestServiceImpl service_; }; template class CommonStressTestSyncServerLowThreadCount : public BaseClass { public: void SetUp() override { ServerBuilder builder; ResourceQuota quota; this->SetUpStart(&builder, &service_); quota.SetMaxThreads(4); builder.SetResourceQuota(quota); this->SetUpEnd(&builder); } void TearDown() override { this->TearDownStart(); this->TearDownEnd(); } private: TestServiceImpl service_; }; template class CommonStressTestAsyncServer : public BaseClass { public: CommonStressTestAsyncServer() : contexts_(kNumAsyncServerThreads * 100) {} void SetUp() override { shutting_down_ = false; ServerBuilder builder; this->SetUpStart(&builder, &service_); cq_ = builder.AddCompletionQueue(); this->SetUpEnd(&builder); for (int i = 0; i < kNumAsyncServerThreads * 100; i++) { RefreshContext(i); } for (int i = 0; i < kNumAsyncServerThreads; i++) { server_threads_.emplace_back(&CommonStressTestAsyncServer::ProcessRpcs, this); } } void TearDown() override { { grpc::internal::MutexLock l(&mu_); this->TearDownStart(); shutting_down_ = true; cq_->Shutdown(); } for (int i = 0; i < kNumAsyncServerThreads; i++) { server_threads_[i].join(); } void* ignored_tag; bool ignored_ok; while (cq_->Next(&ignored_tag, &ignored_ok)) { } this->TearDownEnd(); } private: void ProcessRpcs() { void* tag; bool ok; while (cq_->Next(&tag, &ok)) { if (ok) { int i = static_cast(reinterpret_cast(tag)); switch (contexts_[i].state) { case Context::READY: { contexts_[i].state = Context::DONE; EchoResponse send_response; send_response.set_message(contexts_[i].recv_request.message()); contexts_[i].response_writer->Finish(send_response, Status::OK, tag); break; } case Context::DONE: RefreshContext(i); break; } } } } void RefreshContext(int i) { grpc::internal::MutexLock l(&mu_); if (!shutting_down_) { contexts_[i].state = Context::READY; contexts_[i].srv_ctx.reset(new ServerContext); contexts_[i].response_writer.reset( new grpc::ServerAsyncResponseWriter( contexts_[i].srv_ctx.get())); service_.RequestEcho(contexts_[i].srv_ctx.get(), &contexts_[i].recv_request, contexts_[i].response_writer.get(), cq_.get(), cq_.get(), reinterpret_cast(i)); } } struct Context { std::unique_ptr srv_ctx; std::unique_ptr> response_writer; EchoRequest recv_request; enum { READY, DONE } state; }; std::vector contexts_; ::grpc::testing::EchoTestService::AsyncService service_; std::unique_ptr cq_; bool shutting_down_; grpc::internal::Mutex mu_; std::vector server_threads_; }; template class End2endTest : public ::testing::Test { protected: End2endTest() {} void SetUp() override { common_.SetUp(); } void TearDown() override { common_.TearDown(); } void ResetStub() { common_.ResetStub(); } Common common_; }; static void SendRpc(grpc::testing::EchoTestService::Stub* stub, int num_rpcs, bool allow_exhaustion, gpr_atm* errors) { EchoRequest request; EchoResponse response; request.set_message("Hello"); for (int i = 0; i < num_rpcs; ++i) { ClientContext context; Status s = stub->Echo(&context, request, &response); EXPECT_TRUE(s.ok() || (allow_exhaustion && s.error_code() == StatusCode::RESOURCE_EXHAUSTED)); if (!s.ok()) { if (!(allow_exhaustion && s.error_code() == StatusCode::RESOURCE_EXHAUSTED)) { gpr_log(GPR_ERROR, "RPC error: %d: %s", s.error_code(), s.error_message().c_str()); } gpr_atm_no_barrier_fetch_add(errors, static_cast(1)); } else { EXPECT_EQ(response.message(), request.message()); } } } typedef ::testing::Types< CommonStressTestSyncServer>, CommonStressTestSyncServer>, CommonStressTestSyncServerLowThreadCount< CommonStressTestInproc>, CommonStressTestAsyncServer< CommonStressTestInsecure>, CommonStressTestAsyncServer>> CommonTypes; TYPED_TEST_SUITE(End2endTest, CommonTypes); TYPED_TEST(End2endTest, ThreadStress) { this->common_.ResetStub(); std::vector threads; gpr_atm errors; gpr_atm_rel_store(&errors, static_cast(0)); threads.reserve(kNumThreads); for (int i = 0; i < kNumThreads; ++i) { threads.emplace_back(SendRpc, this->common_.GetStub(), kNumRpcs, this->common_.AllowExhaustion(), &errors); } for (int i = 0; i < kNumThreads; ++i) { threads[i].join(); } uint64_t error_cnt = static_cast(gpr_atm_no_barrier_load(&errors)); if (error_cnt != 0) { gpr_log(GPR_INFO, "RPC error count: %" PRIu64, error_cnt); } // If this test allows resource exhaustion, expect that it actually sees some if (this->common_.AllowExhaustion()) { EXPECT_GT(error_cnt, static_cast(0)); } } template class AsyncClientEnd2endTest : public ::testing::Test { protected: AsyncClientEnd2endTest() : rpcs_outstanding_(0) {} void SetUp() override { common_.SetUp(); } void TearDown() override { void* ignored_tag; bool ignored_ok; while (cq_.Next(&ignored_tag, &ignored_ok)) { } common_.TearDown(); } void Wait() { grpc::internal::MutexLock l(&mu_); while (rpcs_outstanding_ != 0) { cv_.Wait(&mu_); } cq_.Shutdown(); } struct AsyncClientCall { EchoResponse response; ClientContext context; Status status; std::unique_ptr> response_reader; }; void AsyncSendRpc(int num_rpcs) { for (int i = 0; i < num_rpcs; ++i) { AsyncClientCall* call = new AsyncClientCall; EchoRequest request; request.set_message("Hello: " + std::to_string(i)); call->response_reader = common_.GetStub()->AsyncEcho(&call->context, request, &cq_); call->response_reader->Finish(&call->response, &call->status, call); grpc::internal::MutexLock l(&mu_); rpcs_outstanding_++; } } void AsyncCompleteRpc() { while (true) { void* got_tag; bool ok = false; if (!cq_.Next(&got_tag, &ok)) break; AsyncClientCall* call = static_cast(got_tag); if (!ok) { gpr_log(GPR_DEBUG, "Error: %d", call->status.error_code()); } delete call; bool notify; { grpc::internal::MutexLock l(&mu_); rpcs_outstanding_--; notify = (rpcs_outstanding_ == 0); } if (notify) { cv_.Signal(); } } } Common common_; CompletionQueue cq_; grpc::internal::Mutex mu_; grpc::internal::CondVar cv_; int rpcs_outstanding_; }; TYPED_TEST_SUITE(AsyncClientEnd2endTest, CommonTypes); TYPED_TEST(AsyncClientEnd2endTest, ThreadStress) { this->common_.ResetStub(); std::vector send_threads, completion_threads; for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { completion_threads.emplace_back( &AsyncClientEnd2endTest_ThreadStress_Test::AsyncCompleteRpc, this); } for (int i = 0; i < kNumAsyncSendThreads; ++i) { send_threads.emplace_back( &AsyncClientEnd2endTest_ThreadStress_Test::AsyncSendRpc, this, kNumRpcs); } for (int i = 0; i < kNumAsyncSendThreads; ++i) { send_threads[i].join(); } this->Wait(); for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { completion_threads[i].join(); } } } // namespace testing } // namespace grpc int main(int argc, char** argv) { grpc::testing::TestEnvironment env(argc, argv); ::testing::InitGoogleTest(&argc, argv); return RUN_ALL_TESTS(); }