/* * * Copyright 2018 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */ #include #include #include #include #include #include #include #include "absl/memory/memory.h" #include #include #include #include #include #include #include #include #include #include "src/core/lib/gpr/env.h" #include "src/core/lib/iomgr/iomgr.h" #include "src/proto/grpc/testing/echo.grpc.pb.h" #include "test/core/util/port.h" #include "test/core/util/test_config.h" #include "test/cpp/end2end/interceptors_util.h" #include "test/cpp/end2end/test_service_impl.h" #include "test/cpp/util/byte_buffer_proto_helper.h" #include "test/cpp/util/string_ref_helper.h" #include "test/cpp/util/test_credentials_provider.h" namespace grpc { namespace testing { namespace { enum class Protocol { INPROC, TCP }; class TestScenario { public: TestScenario(bool serve_callback, Protocol protocol, bool intercept, const std::string& creds_type) : callback_server(serve_callback), protocol(protocol), use_interceptors(intercept), credentials_type(creds_type) {} void Log() const; bool callback_server; Protocol protocol; bool use_interceptors; const std::string credentials_type; }; static std::ostream& operator<<(std::ostream& out, const TestScenario& scenario) { return out << "TestScenario{callback_server=" << (scenario.callback_server ? "true" : "false") << ",protocol=" << (scenario.protocol == Protocol::INPROC ? "INPROC" : "TCP") << ",intercept=" << (scenario.use_interceptors ? "true" : "false") << ",creds=" << scenario.credentials_type << "}"; } void TestScenario::Log() const { std::ostringstream out; out << *this; gpr_log(GPR_DEBUG, "%s", out.str().c_str()); } class ClientCallbackEnd2endTest : public ::testing::TestWithParam { protected: ClientCallbackEnd2endTest() { GetParam().Log(); } void SetUp() override { ServerBuilder builder; auto server_creds = GetCredentialsProvider()->GetServerCredentials( GetParam().credentials_type); // TODO(vjpai): Support testing of AuthMetadataProcessor if (GetParam().protocol == Protocol::TCP) { picked_port_ = grpc_pick_unused_port_or_die(); server_address_ << "localhost:" << picked_port_; builder.AddListeningPort(server_address_.str(), server_creds); } if (!GetParam().callback_server) { builder.RegisterService(&service_); } else { builder.RegisterService(&callback_service_); } if (GetParam().use_interceptors) { std::vector< std::unique_ptr> creators; // Add 20 phony server interceptors creators.reserve(20); for (auto i = 0; i < 20; i++) { creators.push_back(absl::make_unique()); } builder.experimental().SetInterceptorCreators(std::move(creators)); } server_ = builder.BuildAndStart(); is_server_started_ = true; } void ResetStub( std::unique_ptr interceptor = nullptr) { ChannelArguments args; auto channel_creds = GetCredentialsProvider()->GetChannelCredentials( GetParam().credentials_type, &args); auto interceptors = CreatePhonyClientInterceptors(); if (interceptor != nullptr) interceptors.push_back(std::move(interceptor)); switch (GetParam().protocol) { case Protocol::TCP: if (!GetParam().use_interceptors) { channel_ = ::grpc::CreateCustomChannel(server_address_.str(), channel_creds, args); } else { channel_ = CreateCustomChannelWithInterceptors( server_address_.str(), channel_creds, args, std::move(interceptors)); } break; case Protocol::INPROC: if (!GetParam().use_interceptors) { channel_ = server_->InProcessChannel(args); } else { channel_ = server_->experimental().InProcessChannelWithInterceptors( args, std::move(interceptors)); } break; default: assert(false); } stub_ = grpc::testing::EchoTestService::NewStub(channel_); generic_stub_ = absl::make_unique(channel_); PhonyInterceptor::Reset(); } void TearDown() override { if (is_server_started_) { // Although we would normally do an explicit shutdown, the server // should also work correctly with just a destructor call. The regular // end2end test uses explicit shutdown, so let this one just do reset. server_.reset(); } if (picked_port_ > 0) { grpc_recycle_unused_port(picked_port_); } } void SendRpcs(int num_rpcs, bool with_binary_metadata) { std::string test_string(""); for (int i = 0; i < num_rpcs; i++) { EchoRequest request; EchoResponse response; ClientContext cli_ctx; test_string += "Hello world. "; request.set_message(test_string); std::string val; if (with_binary_metadata) { request.mutable_param()->set_echo_metadata(true); char bytes[8] = {'\0', '\1', '\2', '\3', '\4', '\5', '\6', static_cast(i)}; val = std::string(bytes, 8); cli_ctx.AddMetadata("custom-bin", val); } cli_ctx.set_compression_algorithm(GRPC_COMPRESS_GZIP); std::mutex mu; std::condition_variable cv; bool done = false; stub_->async()->Echo( &cli_ctx, &request, &response, [&cli_ctx, &request, &response, &done, &mu, &cv, val, with_binary_metadata](Status s) { GPR_ASSERT(s.ok()); EXPECT_EQ(request.message(), response.message()); if (with_binary_metadata) { EXPECT_EQ( 1u, cli_ctx.GetServerTrailingMetadata().count("custom-bin")); EXPECT_EQ(val, ToString(cli_ctx.GetServerTrailingMetadata() .find("custom-bin") ->second)); } std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } } void SendRpcsGeneric(int num_rpcs, bool maybe_except, const char* suffix_for_stats) { const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); std::string test_string(""); for (int i = 0; i < num_rpcs; i++) { EchoRequest request; std::unique_ptr send_buf; ByteBuffer recv_buf; ClientContext cli_ctx; test_string += "Hello world. "; request.set_message(test_string); send_buf = SerializeToByteBuffer(&request); std::mutex mu; std::condition_variable cv; bool done = false; StubOptions options(suffix_for_stats); generic_stub_->UnaryCall( &cli_ctx, kMethodName, options, send_buf.get(), &recv_buf, [&request, &recv_buf, &done, &mu, &cv, maybe_except](Status s) { GPR_ASSERT(s.ok()); EchoResponse response; EXPECT_TRUE(ParseFromByteBuffer(&recv_buf, &response)); EXPECT_EQ(request.message(), response.message()); std::lock_guard l(mu); done = true; cv.notify_one(); #if GRPC_ALLOW_EXCEPTIONS if (maybe_except) { throw -1; } #else GPR_ASSERT(!maybe_except); #endif }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } } void SendGenericEchoAsBidi(int num_rpcs, int reuses, bool do_writes_done, const char* suffix_for_stats) { const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); std::string test_string(""); for (int i = 0; i < num_rpcs; i++) { test_string += "Hello world. "; class Client : public grpc::ClientBidiReactor { public: Client(ClientCallbackEnd2endTest* test, const std::string& method_name, const char* suffix_for_stats, const std::string& test_str, int reuses, bool do_writes_done) : reuses_remaining_(reuses), do_writes_done_(do_writes_done) { activate_ = [this, test, method_name, suffix_for_stats, test_str] { if (reuses_remaining_ > 0) { cli_ctx_ = absl::make_unique(); reuses_remaining_--; StubOptions options(suffix_for_stats); test->generic_stub_->PrepareBidiStreamingCall( cli_ctx_.get(), method_name, options, this); request_.set_message(test_str); send_buf_ = SerializeToByteBuffer(&request_); StartWrite(send_buf_.get()); StartRead(&recv_buf_); StartCall(); } else { std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } }; activate_(); } void OnWriteDone(bool /*ok*/) override { if (do_writes_done_) { StartWritesDone(); } } void OnReadDone(bool /*ok*/) override { EchoResponse response; EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response)); EXPECT_EQ(request_.message(), response.message()); }; void OnDone(const Status& s) override { EXPECT_TRUE(s.ok()); activate_(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } EchoRequest request_; std::unique_ptr send_buf_; ByteBuffer recv_buf_; std::unique_ptr cli_ctx_; int reuses_remaining_; std::function activate_; std::mutex mu_; std::condition_variable cv_; bool done_ = false; const bool do_writes_done_; }; Client rpc(this, kMethodName, suffix_for_stats, test_string, reuses, do_writes_done); rpc.Await(); } } bool is_server_started_{false}; int picked_port_{0}; std::shared_ptr channel_; std::unique_ptr stub_; std::unique_ptr generic_stub_; TestServiceImpl service_; CallbackTestServiceImpl callback_service_; std::unique_ptr server_; std::ostringstream server_address_; }; TEST_P(ClientCallbackEnd2endTest, SimpleRpc) { ResetStub(); SendRpcs(1, false); } TEST_P(ClientCallbackEnd2endTest, SimpleRpcExpectedError) { ResetStub(); EchoRequest request; EchoResponse response; ClientContext cli_ctx; ErrorStatus error_status; request.set_message("Hello failure"); error_status.set_code(1); // CANCELLED error_status.set_error_message("cancel error message"); *request.mutable_param()->mutable_expected_error() = error_status; std::mutex mu; std::condition_variable cv; bool done = false; stub_->async()->Echo(&cli_ctx, &request, &response, [&response, &done, &mu, &cv, &error_status](Status s) { EXPECT_EQ("", response.message()); EXPECT_EQ(error_status.code(), s.error_code()); EXPECT_EQ(error_status.error_message(), s.error_message()); std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLockNested) { ResetStub(); // The request/response state associated with an RPC and the synchronization // variables needed to notify its completion. struct RpcState { std::mutex mu; std::condition_variable cv; bool done = false; EchoRequest request; EchoResponse response; ClientContext cli_ctx; RpcState() = default; ~RpcState() { // Grab the lock to prevent destruction while another is still holding // lock std::lock_guard lock(mu); } }; std::vector rpc_state(3); for (size_t i = 0; i < rpc_state.size(); i++) { std::string message = "Hello locked world"; message += std::to_string(i); rpc_state[i].request.set_message(message); } // Grab a lock and then start an RPC whose callback grabs the same lock and // then calls this function to start the next RPC under lock (up to a limit of // the size of the rpc_state vector). std::function nested_call = [this, &nested_call, &rpc_state](int index) { std::lock_guard l(rpc_state[index].mu); stub_->async()->Echo(&rpc_state[index].cli_ctx, &rpc_state[index].request, &rpc_state[index].response, [index, &nested_call, &rpc_state](Status s) { std::lock_guard l1(rpc_state[index].mu); EXPECT_TRUE(s.ok()); rpc_state[index].done = true; rpc_state[index].cv.notify_all(); // Call the next level of nesting if possible if (index + 1 < int(rpc_state.size())) { nested_call(index + 1); } }); }; nested_call(0); // Wait for completion notifications from all RPCs. Order doesn't matter. for (RpcState& state : rpc_state) { std::unique_lock l(state.mu); while (!state.done) { state.cv.wait(l); } EXPECT_EQ(state.request.message(), state.response.message()); } } TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLock) { ResetStub(); std::mutex mu; std::condition_variable cv; bool done = false; EchoRequest request; request.set_message("Hello locked world."); EchoResponse response; ClientContext cli_ctx; { std::lock_guard l(mu); stub_->async()->Echo(&cli_ctx, &request, &response, [&mu, &cv, &done, &request, &response](Status s) { std::lock_guard l(mu); EXPECT_TRUE(s.ok()); EXPECT_EQ(request.message(), response.message()); done = true; cv.notify_one(); }); } std::unique_lock l(mu); while (!done) { cv.wait(l); } } TEST_P(ClientCallbackEnd2endTest, SequentialRpcs) { ResetStub(); SendRpcs(10, false); } TEST_P(ClientCallbackEnd2endTest, SendClientInitialMetadata) { ResetStub(); SimpleRequest request; SimpleResponse response; ClientContext cli_ctx; cli_ctx.AddMetadata(kCheckClientInitialMetadataKey, kCheckClientInitialMetadataVal); std::mutex mu; std::condition_variable cv; bool done = false; stub_->async()->CheckClientInitialMetadata( &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) { GPR_ASSERT(s.ok()); std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } TEST_P(ClientCallbackEnd2endTest, SimpleRpcWithBinaryMetadata) { ResetStub(); SendRpcs(1, true); } TEST_P(ClientCallbackEnd2endTest, SequentialRpcsWithVariedBinaryMetadataValue) { ResetStub(); SendRpcs(10, true); } TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcs) { ResetStub(absl::make_unique( "/grpc.testing.EchoTestService/Echo", nullptr)); SendRpcsGeneric(10, false, /*suffix_for_stats=*/nullptr); } TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsWithSuffix) { ResetStub(absl::make_unique( "/grpc.testing.EchoTestService/Echo", "TestSuffix")); SendRpcsGeneric(10, false, "TestSuffix"); } TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidi) { ResetStub(absl::make_unique( "/grpc.testing.EchoTestService/Echo", nullptr)); SendGenericEchoAsBidi(10, 1, /*do_writes_done=*/true, /*suffix_for_stats=*/nullptr); } TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidiWithSuffix) { ResetStub(absl::make_unique( "/grpc.testing.EchoTestService/Echo", "TestSuffix")); SendGenericEchoAsBidi(10, 1, /*do_writes_done=*/true, "TestSuffix"); } TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidiWithReactorReuse) { ResetStub(); SendGenericEchoAsBidi(10, 10, /*do_writes_done=*/true, /*suffix_for_stats=*/nullptr); } TEST_P(ClientCallbackEnd2endTest, GenericRpcNoWritesDone) { ResetStub(); SendGenericEchoAsBidi(1, 1, /*do_writes_done=*/false, /*suffix_for_stats=*/nullptr); } #if GRPC_ALLOW_EXCEPTIONS TEST_P(ClientCallbackEnd2endTest, ExceptingRpc) { ResetStub(); SendRpcsGeneric(10, true, nullptr); } #endif TEST_P(ClientCallbackEnd2endTest, MultipleRpcsWithVariedBinaryMetadataValue) { ResetStub(); std::vector threads; threads.reserve(10); for (int i = 0; i < 10; ++i) { threads.emplace_back([this] { SendRpcs(10, true); }); } for (int i = 0; i < 10; ++i) { threads[i].join(); } } TEST_P(ClientCallbackEnd2endTest, MultipleRpcs) { ResetStub(); std::vector threads; threads.reserve(10); for (int i = 0; i < 10; ++i) { threads.emplace_back([this] { SendRpcs(10, false); }); } for (int i = 0; i < 10; ++i) { threads[i].join(); } } TEST_P(ClientCallbackEnd2endTest, CancelRpcBeforeStart) { ResetStub(); EchoRequest request; EchoResponse response; ClientContext context; request.set_message("hello"); context.TryCancel(); std::mutex mu; std::condition_variable cv; bool done = false; stub_->async()->Echo(&context, &request, &response, [&response, &done, &mu, &cv](Status s) { EXPECT_EQ("", response.message()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, RequestEchoServerCancel) { ResetStub(); EchoRequest request; EchoResponse response; ClientContext context; request.set_message("hello"); context.AddMetadata(kServerTryCancelRequest, std::to_string(CANCEL_BEFORE_PROCESSING)); std::mutex mu; std::condition_variable cv; bool done = false; stub_->async()->Echo(&context, &request, &response, [&done, &mu, &cv](Status s) { EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } struct ClientCancelInfo { bool cancel{false}; int ops_before_cancel; ClientCancelInfo() : cancel{false} {} explicit ClientCancelInfo(int ops) : cancel{true}, ops_before_cancel{ops} {} }; class WriteClient : public grpc::ClientWriteReactor { public: WriteClient(grpc::testing::EchoTestService::Stub* stub, ServerTryCancelRequestPhase server_try_cancel, int num_msgs_to_send, ClientCancelInfo client_cancel = {}) : server_try_cancel_(server_try_cancel), num_msgs_to_send_(num_msgs_to_send), client_cancel_{client_cancel} { std::string msg{"Hello server."}; for (int i = 0; i < num_msgs_to_send; i++) { desired_ += msg; } if (server_try_cancel != DO_NOT_CANCEL) { // Send server_try_cancel value in the client metadata context_.AddMetadata(kServerTryCancelRequest, std::to_string(server_try_cancel)); } context_.set_initial_metadata_corked(true); stub->async()->RequestStream(&context_, &response_, this); StartCall(); request_.set_message(msg); MaybeWrite(); } void OnWriteDone(bool ok) override { if (ok) { num_msgs_sent_++; MaybeWrite(); } } void OnDone(const Status& s) override { gpr_log(GPR_INFO, "Sent %d messages", num_msgs_sent_); int num_to_send = (client_cancel_.cancel) ? std::min(num_msgs_to_send_, client_cancel_.ops_before_cancel) : num_msgs_to_send_; switch (server_try_cancel_) { case CANCEL_BEFORE_PROCESSING: case CANCEL_DURING_PROCESSING: // If the RPC is canceled by server before / during messages from the // client, it means that the client most likely did not get a chance to // send all the messages it wanted to send. i.e num_msgs_sent <= // num_msgs_to_send EXPECT_LE(num_msgs_sent_, num_to_send); break; case DO_NOT_CANCEL: case CANCEL_AFTER_PROCESSING: // If the RPC was not canceled or canceled after all messages were read // by the server, the client did get a chance to send all its messages EXPECT_EQ(num_msgs_sent_, num_to_send); break; default: assert(false); break; } if ((server_try_cancel_ == DO_NOT_CANCEL) && !client_cancel_.cancel) { EXPECT_TRUE(s.ok()); EXPECT_EQ(response_.message(), desired_); } else { EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); } std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: void MaybeWrite() { if (client_cancel_.cancel && num_msgs_sent_ == client_cancel_.ops_before_cancel) { context_.TryCancel(); } else if (num_msgs_to_send_ > num_msgs_sent_ + 1) { StartWrite(&request_); } else if (num_msgs_to_send_ == num_msgs_sent_ + 1) { StartWriteLast(&request_, WriteOptions()); } } EchoRequest request_; EchoResponse response_; ClientContext context_; const ServerTryCancelRequestPhase server_try_cancel_; int num_msgs_sent_{0}; const int num_msgs_to_send_; std::string desired_; const ClientCancelInfo client_cancel_; std::mutex mu_; std::condition_variable cv_; bool done_ = false; }; TEST_P(ClientCallbackEnd2endTest, RequestStream) { ResetStub(); WriteClient test{stub_.get(), DO_NOT_CANCEL, 3}; test.Await(); // Make sure that the server interceptors were not notified to cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, ClientCancelsRequestStream) { ResetStub(); WriteClient test{stub_.get(), DO_NOT_CANCEL, 3, ClientCancelInfo{2}}; test.Await(); // Make sure that the server interceptors got the cancel if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel before doing reading the request TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelBeforeReads) { ResetStub(); WriteClient test{stub_.get(), CANCEL_BEFORE_PROCESSING, 1}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel while reading a request from the stream in parallel TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelDuringRead) { ResetStub(); WriteClient test{stub_.get(), CANCEL_DURING_PROCESSING, 10}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel after reading all the requests but before returning to the // client TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelAfterReads) { ResetStub(); WriteClient test{stub_.get(), CANCEL_AFTER_PROCESSING, 4}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, UnaryReactor) { ResetStub(); class UnaryClient : public grpc::ClientUnaryReactor { public: explicit UnaryClient(grpc::testing::EchoTestService::Stub* stub) { cli_ctx_.AddMetadata("key1", "val1"); cli_ctx_.AddMetadata("key2", "val2"); request_.mutable_param()->set_echo_metadata_initially(true); request_.set_message("Hello metadata"); stub->async()->Echo(&cli_ctx_, &request_, &response_, this); StartCall(); } void OnReadInitialMetadataDone(bool ok) override { EXPECT_TRUE(ok); EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1")); EXPECT_EQ( "val1", ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second)); EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2")); EXPECT_EQ( "val2", ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second)); initial_metadata_done_ = true; } void OnDone(const Status& s) override { EXPECT_TRUE(initial_metadata_done_); EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size()); EXPECT_TRUE(s.ok()); EXPECT_EQ(request_.message(), response_.message()); std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: EchoRequest request_; EchoResponse response_; ClientContext cli_ctx_; std::mutex mu_; std::condition_variable cv_; bool done_{false}; bool initial_metadata_done_{false}; }; UnaryClient test{stub_.get()}; test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, GenericUnaryReactor) { const std::string kMethodName("/grpc.testing.EchoTestService/Echo"); constexpr char kSuffixForStats[] = "TestSuffixForStats"; ResetStub( absl::make_unique(kMethodName, kSuffixForStats)); class UnaryClient : public grpc::ClientUnaryReactor { public: UnaryClient(grpc::GenericStub* stub, const std::string& method_name, const char* suffix_for_stats) { cli_ctx_.AddMetadata("key1", "val1"); cli_ctx_.AddMetadata("key2", "val2"); request_.mutable_param()->set_echo_metadata_initially(true); request_.set_message("Hello metadata"); send_buf_ = SerializeToByteBuffer(&request_); StubOptions options(suffix_for_stats); stub->PrepareUnaryCall(&cli_ctx_, method_name, options, send_buf_.get(), &recv_buf_, this); StartCall(); } void OnReadInitialMetadataDone(bool ok) override { EXPECT_TRUE(ok); EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1")); EXPECT_EQ( "val1", ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second)); EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2")); EXPECT_EQ( "val2", ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second)); initial_metadata_done_ = true; } void OnDone(const Status& s) override { EXPECT_TRUE(initial_metadata_done_); EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size()); EXPECT_TRUE(s.ok()); EchoResponse response; EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response)); EXPECT_EQ(request_.message(), response.message()); std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: EchoRequest request_; std::unique_ptr send_buf_; ByteBuffer recv_buf_; ClientContext cli_ctx_; std::mutex mu_; std::condition_variable cv_; bool done_{false}; bool initial_metadata_done_{false}; }; UnaryClient test{generic_stub_.get(), kMethodName, kSuffixForStats}; test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } class ReadClient : public grpc::ClientReadReactor { public: ReadClient(grpc::testing::EchoTestService::Stub* stub, ServerTryCancelRequestPhase server_try_cancel, ClientCancelInfo client_cancel = {}) : server_try_cancel_(server_try_cancel), client_cancel_{client_cancel} { if (server_try_cancel_ != DO_NOT_CANCEL) { // Send server_try_cancel value in the client metadata context_.AddMetadata(kServerTryCancelRequest, std::to_string(server_try_cancel)); } request_.set_message("Hello client "); stub->async()->ResponseStream(&context_, &request_, this); if (client_cancel_.cancel && reads_complete_ == client_cancel_.ops_before_cancel) { context_.TryCancel(); } // Even if we cancel, read until failure because there might be responses // pending StartRead(&response_); StartCall(); } void OnReadDone(bool ok) override { if (!ok) { if (server_try_cancel_ == DO_NOT_CANCEL && !client_cancel_.cancel) { EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend); } } else { EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); EXPECT_EQ(response_.message(), request_.message() + std::to_string(reads_complete_)); reads_complete_++; if (client_cancel_.cancel && reads_complete_ == client_cancel_.ops_before_cancel) { context_.TryCancel(); } // Even if we cancel, read until failure because there might be responses // pending StartRead(&response_); } } void OnDone(const Status& s) override { gpr_log(GPR_INFO, "Read %d messages", reads_complete_); switch (server_try_cancel_) { case DO_NOT_CANCEL: if (!client_cancel_.cancel || client_cancel_.ops_before_cancel > kServerDefaultResponseStreamsToSend) { EXPECT_TRUE(s.ok()); EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend); } else { EXPECT_GE(reads_complete_, client_cancel_.ops_before_cancel); EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); // Status might be ok or cancelled depending on whether server // sent status before client cancel went through if (!s.ok()) { EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); } } break; case CANCEL_BEFORE_PROCESSING: EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); EXPECT_EQ(reads_complete_, 0); break; case CANCEL_DURING_PROCESSING: case CANCEL_AFTER_PROCESSING: // If server canceled while writing messages, client must have read // less than or equal to the expected number of messages. Even if the // server canceled after writing all messages, the RPC may be canceled // before the Client got a chance to read all the messages. EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend); break; default: assert(false); } std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: EchoRequest request_; EchoResponse response_; ClientContext context_; const ServerTryCancelRequestPhase server_try_cancel_; int reads_complete_{0}; const ClientCancelInfo client_cancel_; std::mutex mu_; std::condition_variable cv_; bool done_ = false; }; TEST_P(ClientCallbackEnd2endTest, ResponseStream) { ResetStub(); ReadClient test{stub_.get(), DO_NOT_CANCEL}; test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, ClientCancelsResponseStream) { ResetStub(); ReadClient test{stub_.get(), DO_NOT_CANCEL, ClientCancelInfo{2}}; test.Await(); // Because cancel in this case races with server finish, we can't be sure that // server interceptors even see cancellation } // Server to cancel before sending any response messages TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelBefore) { ResetStub(); ReadClient test{stub_.get(), CANCEL_BEFORE_PROCESSING}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel while writing a response to the stream in parallel TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelDuring) { ResetStub(); ReadClient test{stub_.get(), CANCEL_DURING_PROCESSING}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel after writing all the respones to the stream but before // returning to the client TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelAfter) { ResetStub(); ReadClient test{stub_.get(), CANCEL_AFTER_PROCESSING}; test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } class BidiClient : public grpc::ClientBidiReactor { public: BidiClient(grpc::testing::EchoTestService::Stub* stub, ServerTryCancelRequestPhase server_try_cancel, int num_msgs_to_send, bool cork_metadata, bool first_write_async, ClientCancelInfo client_cancel = {}) : server_try_cancel_(server_try_cancel), msgs_to_send_{num_msgs_to_send}, client_cancel_{client_cancel} { if (server_try_cancel_ != DO_NOT_CANCEL) { // Send server_try_cancel value in the client metadata context_.AddMetadata(kServerTryCancelRequest, std::to_string(server_try_cancel)); } request_.set_message("Hello fren "); context_.set_initial_metadata_corked(cork_metadata); stub->async()->BidiStream(&context_, this); MaybeAsyncWrite(first_write_async); StartRead(&response_); StartCall(); } void OnReadDone(bool ok) override { if (!ok) { if (server_try_cancel_ == DO_NOT_CANCEL) { if (!client_cancel_.cancel) { EXPECT_EQ(reads_complete_, msgs_to_send_); } else { EXPECT_LE(reads_complete_, writes_complete_); } } } else { EXPECT_LE(reads_complete_, msgs_to_send_); EXPECT_EQ(response_.message(), request_.message()); reads_complete_++; StartRead(&response_); } } void OnWriteDone(bool ok) override { if (async_write_thread_.joinable()) { async_write_thread_.join(); RemoveHold(); } if (server_try_cancel_ == DO_NOT_CANCEL) { EXPECT_TRUE(ok); } else if (!ok) { return; } writes_complete_++; MaybeWrite(); } void OnDone(const Status& s) override { gpr_log(GPR_INFO, "Sent %d messages", writes_complete_); gpr_log(GPR_INFO, "Read %d messages", reads_complete_); switch (server_try_cancel_) { case DO_NOT_CANCEL: if (!client_cancel_.cancel || client_cancel_.ops_before_cancel > msgs_to_send_) { EXPECT_TRUE(s.ok()); EXPECT_EQ(writes_complete_, msgs_to_send_); EXPECT_EQ(reads_complete_, writes_complete_); } else { EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); EXPECT_EQ(writes_complete_, client_cancel_.ops_before_cancel); EXPECT_LE(reads_complete_, writes_complete_); } break; case CANCEL_BEFORE_PROCESSING: EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); // The RPC is canceled before the server did any work or returned any // reads, but it's possible that some writes took place first from the // client EXPECT_LE(writes_complete_, msgs_to_send_); EXPECT_EQ(reads_complete_, 0); break; case CANCEL_DURING_PROCESSING: EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); EXPECT_LE(writes_complete_, msgs_to_send_); EXPECT_LE(reads_complete_, writes_complete_); break; case CANCEL_AFTER_PROCESSING: EXPECT_FALSE(s.ok()); EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code()); EXPECT_EQ(writes_complete_, msgs_to_send_); // The Server canceled after reading the last message and after writing // the message to the client. However, the RPC cancellation might have // taken effect before the client actually read the response. EXPECT_LE(reads_complete_, writes_complete_); break; default: assert(false); } std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: void MaybeAsyncWrite(bool first_write_async) { if (first_write_async) { // Make sure that we have a write to issue. // TODO(vjpai): Make this work with 0 writes case as well. assert(msgs_to_send_ >= 1); AddHold(); async_write_thread_ = std::thread([this] { std::unique_lock lock(async_write_thread_mu_); async_write_thread_cv_.wait( lock, [this] { return async_write_thread_start_; }); MaybeWrite(); }); std::lock_guard lock(async_write_thread_mu_); async_write_thread_start_ = true; async_write_thread_cv_.notify_one(); return; } MaybeWrite(); } void MaybeWrite() { if (client_cancel_.cancel && writes_complete_ == client_cancel_.ops_before_cancel) { context_.TryCancel(); } else if (writes_complete_ == msgs_to_send_) { StartWritesDone(); } else { StartWrite(&request_); } } EchoRequest request_; EchoResponse response_; ClientContext context_; const ServerTryCancelRequestPhase server_try_cancel_; int reads_complete_{0}; int writes_complete_{0}; const int msgs_to_send_; const ClientCancelInfo client_cancel_; std::mutex mu_; std::condition_variable cv_; bool done_ = false; std::thread async_write_thread_; bool async_write_thread_start_ = false; std::mutex async_write_thread_mu_; std::condition_variable async_write_thread_cv_; }; TEST_P(ClientCallbackEnd2endTest, BidiStream) { ResetStub(); BidiClient test(stub_.get(), DO_NOT_CANCEL, kServerDefaultResponseStreamsToSend, /*cork_metadata=*/false, /*first_write_async=*/false); test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, BidiStreamFirstWriteAsync) { ResetStub(); BidiClient test(stub_.get(), DO_NOT_CANCEL, kServerDefaultResponseStreamsToSend, /*cork_metadata=*/false, /*first_write_async=*/true); test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, BidiStreamCorked) { ResetStub(); BidiClient test(stub_.get(), DO_NOT_CANCEL, kServerDefaultResponseStreamsToSend, /*cork_metadata=*/true, /*first_write_async=*/false); test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, BidiStreamCorkedFirstWriteAsync) { ResetStub(); BidiClient test(stub_.get(), DO_NOT_CANCEL, kServerDefaultResponseStreamsToSend, /*cork_metadata=*/true, /*first_write_async=*/true); test.Await(); // Make sure that the server interceptors were not notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(0, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, ClientCancelsBidiStream) { ResetStub(); BidiClient test(stub_.get(), DO_NOT_CANCEL, kServerDefaultResponseStreamsToSend, /*cork_metadata=*/false, /*first_write_async=*/false, ClientCancelInfo(2)); test.Await(); // Make sure that the server interceptors were notified of a cancel if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel before reading/writing any requests/responses on the stream TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelBefore) { ResetStub(); BidiClient test(stub_.get(), CANCEL_BEFORE_PROCESSING, /*num_msgs_to_send=*/2, /*cork_metadata=*/false, /*first_write_async=*/false); test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel while reading/writing requests/responses on the stream in // parallel TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelDuring) { ResetStub(); BidiClient test(stub_.get(), CANCEL_DURING_PROCESSING, /*num_msgs_to_send=*/10, /*cork_metadata=*/false, /*first_write_async=*/false); test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } // Server to cancel after reading/writing all requests/responses on the stream // but before returning to the client TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelAfter) { ResetStub(); BidiClient test(stub_.get(), CANCEL_AFTER_PROCESSING, /*num_msgs_to_send=*/5, /*cork_metadata=*/false, /*first_write_async=*/false); test.Await(); // Make sure that the server interceptors were notified if (GetParam().use_interceptors) { EXPECT_EQ(20, PhonyInterceptor::GetNumTimesCancel()); } } TEST_P(ClientCallbackEnd2endTest, SimultaneousReadAndWritesDone) { ResetStub(); class Client : public grpc::ClientBidiReactor { public: explicit Client(grpc::testing::EchoTestService::Stub* stub) { request_.set_message("Hello bidi "); stub->async()->BidiStream(&context_, this); StartWrite(&request_); StartCall(); } void OnReadDone(bool ok) override { EXPECT_TRUE(ok); EXPECT_EQ(response_.message(), request_.message()); } void OnWriteDone(bool ok) override { EXPECT_TRUE(ok); // Now send out the simultaneous Read and WritesDone StartWritesDone(); StartRead(&response_); } void OnDone(const Status& s) override { EXPECT_TRUE(s.ok()); EXPECT_EQ(response_.message(), request_.message()); std::unique_lock l(mu_); done_ = true; cv_.notify_one(); } void Await() { std::unique_lock l(mu_); while (!done_) { cv_.wait(l); } } private: EchoRequest request_; EchoResponse response_; ClientContext context_; std::mutex mu_; std::condition_variable cv_; bool done_ = false; } test{stub_.get()}; test.Await(); } TEST_P(ClientCallbackEnd2endTest, UnimplementedRpc) { ChannelArguments args; const auto& channel_creds = GetCredentialsProvider()->GetChannelCredentials( GetParam().credentials_type, &args); std::shared_ptr channel = (GetParam().protocol == Protocol::TCP) ? ::grpc::CreateCustomChannel(server_address_.str(), channel_creds, args) : server_->InProcessChannel(args); std::unique_ptr stub; stub = grpc::testing::UnimplementedEchoService::NewStub(channel); EchoRequest request; EchoResponse response; ClientContext cli_ctx; request.set_message("Hello world."); std::mutex mu; std::condition_variable cv; bool done = false; stub->async()->Unimplemented( &cli_ctx, &request, &response, [&done, &mu, &cv](Status s) { EXPECT_EQ(StatusCode::UNIMPLEMENTED, s.error_code()); EXPECT_EQ("", s.error_message()); std::lock_guard l(mu); done = true; cv.notify_one(); }); std::unique_lock l(mu); while (!done) { cv.wait(l); } } TEST_P(ClientCallbackEnd2endTest, TestTrailersOnlyOnError) { // Note that trailers-only is an HTTP/2 concept so we shouldn't do this test // for any other transport such as inproc. if (GetParam().protocol != Protocol::TCP) { return; } ResetStub(); class Reactor : public grpc::ClientBidiReactor { public: explicit Reactor(grpc::testing::EchoTestService::Stub* stub) { stub->async()->UnimplementedBidi(&context_, this); StartCall(); } void Await() { std::unique_lock l(mu_); while (!done_) { done_cv_.wait(l); } } private: void OnReadInitialMetadataDone(bool ok) override { EXPECT_FALSE(ok); } void OnDone(const Status& s) override { EXPECT_EQ(s.error_code(), grpc::StatusCode::UNIMPLEMENTED); EXPECT_EQ(s.error_message(), ""); std::unique_lock l(mu_); done_ = true; done_cv_.notify_one(); } ClientContext context_; std::mutex mu_; std::condition_variable done_cv_; bool done_ = false; } client(stub_.get()); client.Await(); } TEST_P(ClientCallbackEnd2endTest, ResponseStreamExtraReactionFlowReadsUntilDone) { ResetStub(); class ReadAllIncomingDataClient : public grpc::ClientReadReactor { public: explicit ReadAllIncomingDataClient( grpc::testing::EchoTestService::Stub* stub) { request_.set_message("Hello client "); stub->async()->ResponseStream(&context_, &request_, this); } bool WaitForReadDone() { std::unique_lock l(mu_); while (!read_done_) { read_cv_.wait(l); } read_done_ = false; return read_ok_; } void Await() { std::unique_lock l(mu_); while (!done_) { done_cv_.wait(l); } } // RemoveHold under the same lock used for OnDone to make sure that we don't // call OnDone directly or indirectly from the RemoveHold function. void RemoveHoldUnderLock() { std::unique_lock l(mu_); RemoveHold(); } const Status& status() { std::unique_lock l(mu_); return status_; } private: void OnReadDone(bool ok) override { std::unique_lock l(mu_); read_ok_ = ok; read_done_ = true; read_cv_.notify_one(); } void OnDone(const Status& s) override { std::unique_lock l(mu_); done_ = true; status_ = s; done_cv_.notify_one(); } EchoRequest request_; EchoResponse response_; ClientContext context_; bool read_ok_ = false; bool read_done_ = false; std::mutex mu_; std::condition_variable read_cv_; std::condition_variable done_cv_; bool done_ = false; Status status_; } client{stub_.get()}; int reads_complete = 0; client.AddHold(); client.StartCall(); EchoResponse response; bool read_ok = true; while (read_ok) { client.StartRead(&response); read_ok = client.WaitForReadDone(); if (read_ok) { ++reads_complete; } } client.RemoveHoldUnderLock(); client.Await(); EXPECT_EQ(kServerDefaultResponseStreamsToSend, reads_complete); EXPECT_EQ(client.status().error_code(), grpc::StatusCode::OK); } std::vector CreateTestScenarios(bool test_insecure) { #if TARGET_OS_IPHONE // Workaround Apple CFStream bug gpr_setenv("grpc_cfstream", "0"); #endif std::vector scenarios; std::vector credentials_types{ GetCredentialsProvider()->GetSecureCredentialsTypeList()}; auto insec_ok = [] { // Only allow insecure credentials type when it is registered with the // provider. User may create providers that do not have insecure. return GetCredentialsProvider()->GetChannelCredentials( kInsecureCredentialsType, nullptr) != nullptr; }; if (test_insecure && insec_ok()) { credentials_types.push_back(kInsecureCredentialsType); } GPR_ASSERT(!credentials_types.empty()); bool barr[]{false, true}; Protocol parr[]{Protocol::INPROC, Protocol::TCP}; for (Protocol p : parr) { for (const auto& cred : credentials_types) { // TODO(vjpai): Test inproc with secure credentials when feasible if (p == Protocol::INPROC && (cred != kInsecureCredentialsType || !insec_ok())) { continue; } for (bool callback_server : barr) { for (bool use_interceptors : barr) { scenarios.emplace_back(callback_server, p, use_interceptors, cred); } } } } return scenarios; } INSTANTIATE_TEST_SUITE_P(ClientCallbackEnd2endTest, ClientCallbackEnd2endTest, ::testing::ValuesIn(CreateTestScenarios(true))); } // namespace } // namespace testing } // namespace grpc int main(int argc, char** argv) { ::testing::InitGoogleTest(&argc, argv); grpc::testing::TestEnvironment env(argc, argv); grpc_init(); int ret = RUN_ALL_TESTS(); grpc_shutdown(); return ret; }