mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
471 lines
14 KiB
471 lines
14 KiB
/* |
|
* |
|
* Copyright 2015, Google Inc. |
|
* All rights reserved. |
|
* |
|
* Redistribution and use in source and binary forms, with or without |
|
* modification, are permitted provided that the following conditions are |
|
* met: |
|
* |
|
* * Redistributions of source code must retain the above copyright |
|
* notice, this list of conditions and the following disclaimer. |
|
* * Redistributions in binary form must reproduce the above |
|
* copyright notice, this list of conditions and the following disclaimer |
|
* in the documentation and/or other materials provided with the |
|
* distribution. |
|
* * Neither the name of Google Inc. nor the names of its |
|
* contributors may be used to endorse or promote products derived from |
|
* this software without specific prior written permission. |
|
* |
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
|
* |
|
*/ |
|
|
|
#include <mutex> |
|
#include <thread> |
|
|
|
#include <grpc++/channel.h> |
|
#include <grpc++/client_context.h> |
|
#include <grpc++/create_channel.h> |
|
#include <grpc++/server.h> |
|
#include <grpc++/server_builder.h> |
|
#include <grpc++/server_context.h> |
|
#include <grpc/grpc.h> |
|
#include <grpc/support/thd.h> |
|
#include <grpc/support/time.h> |
|
#include <gtest/gtest.h> |
|
|
|
#include "src/core/lib/surface/api_trace.h" |
|
#include "src/proto/grpc/testing/duplicate/echo_duplicate.grpc.pb.h" |
|
#include "src/proto/grpc/testing/echo.grpc.pb.h" |
|
#include "test/core/util/port.h" |
|
#include "test/core/util/test_config.h" |
|
|
|
using grpc::testing::EchoRequest; |
|
using grpc::testing::EchoResponse; |
|
using std::chrono::system_clock; |
|
|
|
const int kNumThreads = 100; // Number of threads |
|
const int kNumAsyncSendThreads = 2; |
|
const int kNumAsyncReceiveThreads = 50; |
|
const int kNumAsyncServerThreads = 50; |
|
const int kNumRpcs = 1000; // Number of RPCs per thread |
|
|
|
namespace grpc { |
|
namespace testing { |
|
|
|
namespace { |
|
|
|
// When echo_deadline is requested, deadline seen in the ServerContext is set in |
|
// the response in seconds. |
|
void MaybeEchoDeadline(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) { |
|
if (request->has_param() && request->param().echo_deadline()) { |
|
gpr_timespec deadline = gpr_inf_future(GPR_CLOCK_REALTIME); |
|
if (context->deadline() != system_clock::time_point::max()) { |
|
Timepoint2Timespec(context->deadline(), &deadline); |
|
} |
|
response->mutable_param()->set_request_deadline(deadline.tv_sec); |
|
} |
|
} |
|
|
|
} // namespace |
|
|
|
class TestServiceImpl : public ::grpc::testing::EchoTestService::Service { |
|
public: |
|
TestServiceImpl() : signal_client_(false) {} |
|
|
|
Status Echo(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) GRPC_OVERRIDE { |
|
response->set_message(request->message()); |
|
MaybeEchoDeadline(context, request, response); |
|
if (request->has_param() && request->param().client_cancel_after_us()) { |
|
{ |
|
unique_lock<mutex> lock(mu_); |
|
signal_client_ = true; |
|
} |
|
while (!context->IsCancelled()) { |
|
gpr_sleep_until(gpr_time_add( |
|
gpr_now(GPR_CLOCK_REALTIME), |
|
gpr_time_from_micros(request->param().client_cancel_after_us(), |
|
GPR_TIMESPAN))); |
|
} |
|
return Status::CANCELLED; |
|
} else if (request->has_param() && |
|
request->param().server_cancel_after_us()) { |
|
gpr_sleep_until(gpr_time_add( |
|
gpr_now(GPR_CLOCK_REALTIME), |
|
gpr_time_from_micros(request->param().server_cancel_after_us(), |
|
GPR_TIMESPAN))); |
|
return Status::CANCELLED; |
|
} else { |
|
EXPECT_FALSE(context->IsCancelled()); |
|
} |
|
return Status::OK; |
|
} |
|
|
|
// Unimplemented is left unimplemented to test the returned error. |
|
|
|
Status RequestStream(ServerContext* context, |
|
ServerReader<EchoRequest>* reader, |
|
EchoResponse* response) GRPC_OVERRIDE { |
|
EchoRequest request; |
|
response->set_message(""); |
|
while (reader->Read(&request)) { |
|
response->mutable_message()->append(request.message()); |
|
} |
|
return Status::OK; |
|
} |
|
|
|
// Return 3 messages. |
|
// TODO(yangg) make it generic by adding a parameter into EchoRequest |
|
Status ResponseStream(ServerContext* context, const EchoRequest* request, |
|
ServerWriter<EchoResponse>* writer) GRPC_OVERRIDE { |
|
EchoResponse response; |
|
response.set_message(request->message() + "0"); |
|
writer->Write(response); |
|
response.set_message(request->message() + "1"); |
|
writer->Write(response); |
|
response.set_message(request->message() + "2"); |
|
writer->Write(response); |
|
|
|
return Status::OK; |
|
} |
|
|
|
Status BidiStream(ServerContext* context, |
|
ServerReaderWriter<EchoResponse, EchoRequest>* stream) |
|
GRPC_OVERRIDE { |
|
EchoRequest request; |
|
EchoResponse response; |
|
while (stream->Read(&request)) { |
|
gpr_log(GPR_INFO, "recv msg %s", request.message().c_str()); |
|
response.set_message(request.message()); |
|
stream->Write(response); |
|
} |
|
return Status::OK; |
|
} |
|
|
|
bool signal_client() { |
|
unique_lock<mutex> lock(mu_); |
|
return signal_client_; |
|
} |
|
|
|
private: |
|
bool signal_client_; |
|
mutex mu_; |
|
}; |
|
|
|
class TestServiceImplDupPkg |
|
: public ::grpc::testing::duplicate::EchoTestService::Service { |
|
public: |
|
Status Echo(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) GRPC_OVERRIDE { |
|
response->set_message("no package"); |
|
return Status::OK; |
|
} |
|
}; |
|
|
|
template <class Service> |
|
class CommonStressTest { |
|
public: |
|
CommonStressTest() : kMaxMessageSize_(8192) {} |
|
virtual ~CommonStressTest() {} |
|
virtual void SetUp() = 0; |
|
virtual void TearDown() = 0; |
|
void ResetStub() { |
|
std::shared_ptr<Channel> channel = |
|
CreateChannel(server_address_.str(), InsecureChannelCredentials()); |
|
stub_ = grpc::testing::EchoTestService::NewStub(channel); |
|
} |
|
grpc::testing::EchoTestService::Stub* GetStub() { return stub_.get(); } |
|
|
|
protected: |
|
void SetUpStart(ServerBuilder* builder, Service* service) { |
|
int port = grpc_pick_unused_port_or_die(); |
|
server_address_ << "localhost:" << port; |
|
// Setup server |
|
builder->AddListeningPort(server_address_.str(), |
|
InsecureServerCredentials()); |
|
builder->RegisterService(service); |
|
builder->SetMaxMessageSize( |
|
kMaxMessageSize_); // For testing max message size. |
|
builder->RegisterService(&dup_pkg_service_); |
|
} |
|
void SetUpEnd(ServerBuilder* builder) { server_ = builder->BuildAndStart(); } |
|
void TearDownStart() { server_->Shutdown(); } |
|
void TearDownEnd() {} |
|
|
|
private: |
|
std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_; |
|
std::unique_ptr<Server> server_; |
|
std::ostringstream server_address_; |
|
const int kMaxMessageSize_; |
|
TestServiceImplDupPkg dup_pkg_service_; |
|
}; |
|
|
|
class CommonStressTestSyncServer : public CommonStressTest<TestServiceImpl> { |
|
public: |
|
void SetUp() GRPC_OVERRIDE { |
|
ServerBuilder builder; |
|
SetUpStart(&builder, &service_); |
|
SetUpEnd(&builder); |
|
} |
|
void TearDown() GRPC_OVERRIDE { |
|
TearDownStart(); |
|
TearDownEnd(); |
|
} |
|
|
|
private: |
|
TestServiceImpl service_; |
|
}; |
|
|
|
class CommonStressTestAsyncServer |
|
: public CommonStressTest<::grpc::testing::EchoTestService::AsyncService> { |
|
public: |
|
void SetUp() GRPC_OVERRIDE { |
|
shutting_down_ = false; |
|
ServerBuilder builder; |
|
SetUpStart(&builder, &service_); |
|
cq_ = builder.AddCompletionQueue(); |
|
SetUpEnd(&builder); |
|
contexts_ = new Context[kNumAsyncServerThreads * 100]; |
|
for (int i = 0; i < kNumAsyncServerThreads * 100; i++) { |
|
RefreshContext(i); |
|
} |
|
for (int i = 0; i < kNumAsyncServerThreads; i++) { |
|
server_threads_.push_back( |
|
new std::thread(&CommonStressTestAsyncServer::ProcessRpcs, this)); |
|
} |
|
} |
|
void TearDown() GRPC_OVERRIDE { |
|
{ |
|
unique_lock<mutex> l(mu_); |
|
TearDownStart(); |
|
shutting_down_ = true; |
|
cq_->Shutdown(); |
|
} |
|
|
|
for (int i = 0; i < kNumAsyncServerThreads; i++) { |
|
server_threads_[i]->join(); |
|
delete server_threads_[i]; |
|
} |
|
|
|
void* ignored_tag; |
|
bool ignored_ok; |
|
while (cq_->Next(&ignored_tag, &ignored_ok)) |
|
; |
|
TearDownEnd(); |
|
delete[] contexts_; |
|
} |
|
|
|
private: |
|
void ProcessRpcs() { |
|
void* tag; |
|
bool ok; |
|
while (cq_->Next(&tag, &ok)) { |
|
if (ok) { |
|
int i = static_cast<int>(reinterpret_cast<intptr_t>(tag)); |
|
switch (contexts_[i].state) { |
|
case Context::READY: { |
|
contexts_[i].state = Context::DONE; |
|
EchoResponse send_response; |
|
send_response.set_message(contexts_[i].recv_request.message()); |
|
contexts_[i].response_writer->Finish(send_response, Status::OK, |
|
tag); |
|
break; |
|
} |
|
case Context::DONE: |
|
RefreshContext(i); |
|
break; |
|
} |
|
} |
|
} |
|
} |
|
void RefreshContext(int i) { |
|
unique_lock<mutex> l(mu_); |
|
if (!shutting_down_) { |
|
contexts_[i].state = Context::READY; |
|
contexts_[i].srv_ctx.reset(new ServerContext); |
|
contexts_[i].response_writer.reset( |
|
new grpc::ServerAsyncResponseWriter<EchoResponse>( |
|
contexts_[i].srv_ctx.get())); |
|
service_.RequestEcho(contexts_[i].srv_ctx.get(), |
|
&contexts_[i].recv_request, |
|
contexts_[i].response_writer.get(), cq_.get(), |
|
cq_.get(), (void*)(intptr_t)i); |
|
} |
|
} |
|
struct Context { |
|
std::unique_ptr<ServerContext> srv_ctx; |
|
std::unique_ptr<grpc::ServerAsyncResponseWriter<EchoResponse>> |
|
response_writer; |
|
EchoRequest recv_request; |
|
enum { READY, DONE } state; |
|
} * contexts_; |
|
::grpc::testing::EchoTestService::AsyncService service_; |
|
std::unique_ptr<ServerCompletionQueue> cq_; |
|
bool shutting_down_; |
|
mutex mu_; |
|
std::vector<std::thread*> server_threads_; |
|
}; |
|
|
|
template <class Common> |
|
class End2endTest : public ::testing::Test { |
|
protected: |
|
End2endTest() {} |
|
void SetUp() GRPC_OVERRIDE { common_.SetUp(); } |
|
void TearDown() GRPC_OVERRIDE { common_.TearDown(); } |
|
void ResetStub() { common_.ResetStub(); } |
|
|
|
Common common_; |
|
}; |
|
|
|
static void SendRpc(grpc::testing::EchoTestService::Stub* stub, int num_rpcs) { |
|
EchoRequest request; |
|
EchoResponse response; |
|
request.set_message("Hello"); |
|
|
|
for (int i = 0; i < num_rpcs; ++i) { |
|
ClientContext context; |
|
Status s = stub->Echo(&context, request, &response); |
|
EXPECT_EQ(response.message(), request.message()); |
|
EXPECT_TRUE(s.ok()); |
|
} |
|
} |
|
|
|
typedef ::testing::Types<CommonStressTestSyncServer, |
|
CommonStressTestAsyncServer> |
|
CommonTypes; |
|
TYPED_TEST_CASE(End2endTest, CommonTypes); |
|
TYPED_TEST(End2endTest, ThreadStress) { |
|
this->common_.ResetStub(); |
|
std::vector<std::thread*> threads; |
|
for (int i = 0; i < kNumThreads; ++i) { |
|
threads.push_back( |
|
new std::thread(SendRpc, this->common_.GetStub(), kNumRpcs)); |
|
} |
|
for (int i = 0; i < kNumThreads; ++i) { |
|
threads[i]->join(); |
|
delete threads[i]; |
|
} |
|
} |
|
|
|
template <class Common> |
|
class AsyncClientEnd2endTest : public ::testing::Test { |
|
protected: |
|
AsyncClientEnd2endTest() : rpcs_outstanding_(0) {} |
|
|
|
void SetUp() GRPC_OVERRIDE { common_.SetUp(); } |
|
void TearDown() GRPC_OVERRIDE { |
|
void* ignored_tag; |
|
bool ignored_ok; |
|
while (cq_.Next(&ignored_tag, &ignored_ok)) |
|
; |
|
common_.TearDown(); |
|
} |
|
|
|
void Wait() { |
|
unique_lock<mutex> l(mu_); |
|
while (rpcs_outstanding_ != 0) { |
|
cv_.wait(l); |
|
} |
|
|
|
cq_.Shutdown(); |
|
} |
|
|
|
struct AsyncClientCall { |
|
EchoResponse response; |
|
ClientContext context; |
|
Status status; |
|
std::unique_ptr<ClientAsyncResponseReader<EchoResponse>> response_reader; |
|
}; |
|
|
|
void AsyncSendRpc(int num_rpcs) { |
|
for (int i = 0; i < num_rpcs; ++i) { |
|
AsyncClientCall* call = new AsyncClientCall; |
|
EchoRequest request; |
|
request.set_message("Hello: " + std::to_string(i)); |
|
call->response_reader = |
|
common_.GetStub()->AsyncEcho(&call->context, request, &cq_); |
|
call->response_reader->Finish(&call->response, &call->status, |
|
(void*)call); |
|
|
|
unique_lock<mutex> l(mu_); |
|
rpcs_outstanding_++; |
|
} |
|
} |
|
|
|
void AsyncCompleteRpc() { |
|
while (true) { |
|
void* got_tag; |
|
bool ok = false; |
|
if (!cq_.Next(&got_tag, &ok)) break; |
|
AsyncClientCall* call = static_cast<AsyncClientCall*>(got_tag); |
|
if (!ok) { |
|
gpr_log(GPR_DEBUG, "Error: %d", call->status.error_code()); |
|
} |
|
delete call; |
|
|
|
bool notify; |
|
{ |
|
unique_lock<mutex> l(mu_); |
|
rpcs_outstanding_--; |
|
notify = (rpcs_outstanding_ == 0); |
|
} |
|
if (notify) { |
|
cv_.notify_all(); |
|
} |
|
} |
|
} |
|
|
|
Common common_; |
|
CompletionQueue cq_; |
|
mutex mu_; |
|
condition_variable cv_; |
|
int rpcs_outstanding_; |
|
}; |
|
|
|
TYPED_TEST_CASE(AsyncClientEnd2endTest, CommonTypes); |
|
TYPED_TEST(AsyncClientEnd2endTest, ThreadStress) { |
|
this->common_.ResetStub(); |
|
std::vector<std::thread *> send_threads, completion_threads; |
|
for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { |
|
completion_threads.push_back(new std::thread( |
|
&AsyncClientEnd2endTest_ThreadStress_Test<TypeParam>::AsyncCompleteRpc, |
|
this)); |
|
} |
|
for (int i = 0; i < kNumAsyncSendThreads; ++i) { |
|
send_threads.push_back(new std::thread( |
|
&AsyncClientEnd2endTest_ThreadStress_Test<TypeParam>::AsyncSendRpc, |
|
this, kNumRpcs)); |
|
} |
|
for (int i = 0; i < kNumAsyncSendThreads; ++i) { |
|
send_threads[i]->join(); |
|
delete send_threads[i]; |
|
} |
|
|
|
this->Wait(); |
|
for (int i = 0; i < kNumAsyncReceiveThreads; ++i) { |
|
completion_threads[i]->join(); |
|
delete completion_threads[i]; |
|
} |
|
} |
|
|
|
} // namespace testing |
|
} // namespace grpc |
|
|
|
int main(int argc, char** argv) { |
|
grpc_test_init(argc, argv); |
|
::testing::InitGoogleTest(&argc, argv); |
|
return RUN_ALL_TESTS(); |
|
}
|
|
|