mirror of https://github.com/grpc/grpc.git
The C based gRPC (C++, Python, Ruby, Objective-C, PHP, C#)
https://grpc.io/
You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
544 lines
20 KiB
544 lines
20 KiB
// |
|
// |
|
// Copyright 2016 gRPC authors. |
|
// |
|
// Licensed under the Apache License, Version 2.0 (the "License"); |
|
// you may not use this file except in compliance with the License. |
|
// You may obtain a copy of the License at |
|
// |
|
// http://www.apache.org/licenses/LICENSE-2.0 |
|
// |
|
// Unless required by applicable law or agreed to in writing, software |
|
// distributed under the License is distributed on an "AS IS" BASIS, |
|
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. |
|
// See the License for the specific language governing permissions and |
|
// limitations under the License. |
|
// |
|
// |
|
|
|
#ifndef GRPC_TEST_CPP_END2END_TEST_SERVICE_IMPL_H |
|
#define GRPC_TEST_CPP_END2END_TEST_SERVICE_IMPL_H |
|
|
|
#include <condition_variable> |
|
#include <memory> |
|
#include <mutex> |
|
#include <string> |
|
#include <thread> |
|
|
|
#include <gtest/gtest.h> |
|
|
|
#include "absl/log/check.h" |
|
#include "absl/log/log.h" |
|
|
|
#include <grpc/grpc.h> |
|
#include <grpcpp/alarm.h> |
|
#include <grpcpp/security/credentials.h> |
|
#include <grpcpp/server_context.h> |
|
|
|
#include "src/core/lib/gprpp/crash.h" |
|
#include "src/proto/grpc/testing/echo.grpc.pb.h" |
|
#include "test/core/test_util/test_config.h" |
|
#include "test/cpp/util/string_ref_helper.h" |
|
|
|
namespace grpc { |
|
namespace testing { |
|
|
|
const int kServerDefaultResponseStreamsToSend = 3; |
|
const char* const kServerResponseStreamsToSend = "server_responses_to_send"; |
|
const char* const kServerTryCancelRequest = "server_try_cancel"; |
|
const char* const kClientTryCancelRequest = "client_try_cancel"; |
|
const char* const kDebugInfoTrailerKey = "debug-info-bin"; |
|
const char* const kServerFinishAfterNReads = "server_finish_after_n_reads"; |
|
const char* const kServerUseCoalescingApi = "server_use_coalescing_api"; |
|
const char* const kCheckClientInitialMetadataKey = "custom_client_metadata"; |
|
const char* const kCheckClientInitialMetadataVal = "Value for client metadata"; |
|
|
|
typedef enum { |
|
DO_NOT_CANCEL = 0, |
|
CANCEL_BEFORE_PROCESSING, |
|
CANCEL_DURING_PROCESSING, |
|
CANCEL_AFTER_PROCESSING |
|
} ServerTryCancelRequestPhase; |
|
|
|
namespace internal { |
|
// When echo_deadline is requested, deadline seen in the ServerContext is set in |
|
// the response in seconds. |
|
void MaybeEchoDeadline(ServerContextBase* context, const EchoRequest* request, |
|
EchoResponse* response); |
|
|
|
void CheckServerAuthContext(const ServerContextBase* context, |
|
const std::string& expected_transport_security_type, |
|
const std::string& expected_client_identity); |
|
|
|
// Returns the number of pairs in metadata that exactly match the given |
|
// key-value pair. Returns -1 if the pair wasn't found. |
|
int MetadataMatchCount( |
|
const std::multimap<grpc::string_ref, grpc::string_ref>& metadata, |
|
const std::string& key, const std::string& value); |
|
|
|
int GetIntValueFromMetadataHelper( |
|
const char* key, |
|
const std::multimap<grpc::string_ref, grpc::string_ref>& metadata, |
|
int default_value); |
|
|
|
int GetIntValueFromMetadata( |
|
const char* key, |
|
const std::multimap<grpc::string_ref, grpc::string_ref>& metadata, |
|
int default_value); |
|
|
|
void ServerTryCancel(ServerContext* context); |
|
} // namespace internal |
|
|
|
class TestServiceSignaller { |
|
public: |
|
// Waits for at least *desired_rpcs* to to be waiting for a server |
|
// continue notification. |
|
// Returns when *desired_rpcs* reaches that amount, or when we've |
|
// surpassed the timeout, whichever happens first. The return value |
|
// is whatever the number of RPCs waiting for server notification is |
|
// at that time. |
|
int ClientWaitUntilNRpcsStarted(int desired_rpcs, absl::Duration timeout) { |
|
VLOG(2) << "*** enter ClientWaitUntilNRpcsStarted ***"; |
|
absl::Time deadline = absl::Now() + timeout; |
|
std::chrono::system_clock::time_point chrono_deadline = |
|
absl::ToChronoTime(deadline); |
|
std::unique_lock<std::mutex> lock(mu_); |
|
cv_rpc_started_.wait_until(lock, chrono_deadline, [this, desired_rpcs] { |
|
VLOG(2) << "*** desired_rpcs: " << desired_rpcs |
|
<< " rpcs_waiting_for_server_to_continue_: " |
|
<< rpcs_waiting_for_server_to_continue_ << " ***"; |
|
return rpcs_waiting_for_server_to_continue_ >= desired_rpcs; |
|
}); |
|
VLOG(2) << "*** leave ClientWaitUntilNRpcsStarted ***"; |
|
return rpcs_waiting_for_server_to_continue_; |
|
} |
|
void ServerWaitToContinue() { |
|
VLOG(2) << "*** enter ServerWaitToContinue ***"; |
|
std::unique_lock<std::mutex> lock(mu_); |
|
cv_server_continue_.wait(lock, [this] { return server_should_continue_; }); |
|
VLOG(2) << "*** leave ServerWaitToContinue ***"; |
|
} |
|
void SignalClientThatRpcStarted() { |
|
VLOG(2) << "*** SignalClientThatRpcStarted ***"; |
|
std::unique_lock<std::mutex> lock(mu_); |
|
++rpcs_waiting_for_server_to_continue_; |
|
cv_rpc_started_.notify_all(); |
|
} |
|
void SignalServerToContinue() { |
|
VLOG(2) << "*** SignalServerToContinue ***"; |
|
std::unique_lock<std::mutex> lock(mu_); |
|
server_should_continue_ = true; |
|
cv_server_continue_.notify_all(); |
|
} |
|
void Reset() { |
|
std::unique_lock<std::mutex> lock(mu_); |
|
rpcs_waiting_for_server_to_continue_ = 0; |
|
server_should_continue_ = false; |
|
} |
|
|
|
private: |
|
std::mutex mu_; |
|
std::condition_variable cv_rpc_started_; |
|
int rpcs_waiting_for_server_to_continue_ /* GUARDED_BY(mu_) */ = 0; |
|
std::condition_variable cv_server_continue_; |
|
bool server_should_continue_ /* GUARDED_BY(mu_) */ = false; |
|
}; |
|
|
|
template <typename RpcService> |
|
class TestMultipleServiceImpl : public RpcService { |
|
public: |
|
TestMultipleServiceImpl() : signal_client_(false), host_() {} |
|
explicit TestMultipleServiceImpl(const std::string& host) |
|
: signal_client_(false), host_(new std::string(host)) {} |
|
|
|
Status Echo(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) { |
|
if (request->has_param() && |
|
request->param().server_notify_client_when_started()) { |
|
signaller_.SignalClientThatRpcStarted(); |
|
signaller_.ServerWaitToContinue(); |
|
} |
|
|
|
// A bit of sleep to make sure that short deadline tests fail |
|
if (request->has_param() && request->param().server_sleep_us() > 0) { |
|
gpr_sleep_until(gpr_time_add( |
|
gpr_now(GPR_CLOCK_MONOTONIC), |
|
gpr_time_from_micros( |
|
request->param().server_sleep_us() * grpc_test_slowdown_factor(), |
|
GPR_TIMESPAN))); |
|
} |
|
|
|
if (request->has_param() && request->param().server_die()) { |
|
LOG(ERROR) << "The request should not reach application handler."; |
|
CHECK(0); |
|
} |
|
if (request->has_param() && request->param().has_expected_error()) { |
|
const auto& error = request->param().expected_error(); |
|
return Status(static_cast<StatusCode>(error.code()), |
|
error.error_message(), error.binary_error_details()); |
|
} |
|
int server_try_cancel = internal::GetIntValueFromMetadata( |
|
kServerTryCancelRequest, context->client_metadata(), DO_NOT_CANCEL); |
|
if (server_try_cancel > DO_NOT_CANCEL) { |
|
// Since this is a unary RPC, by the time this server handler is called, |
|
// the 'request' message is already read from the client. So the scenarios |
|
// in server_try_cancel don't make much sense. Just cancel the RPC as long |
|
// as server_try_cancel is not DO_NOT_CANCEL |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
response->set_message(request->message()); |
|
internal::MaybeEchoDeadline(context, request, response); |
|
if (host_) { |
|
response->mutable_param()->set_host(*host_); |
|
} else if (request->has_param() && |
|
request->param().echo_host_from_authority_header()) { |
|
auto authority = context->ExperimentalGetAuthority(); |
|
std::string authority_str(authority.data(), authority.size()); |
|
response->mutable_param()->set_host(std::move(authority_str)); |
|
} |
|
if (request->has_param() && request->param().client_cancel_after_us()) { |
|
{ |
|
std::unique_lock<std::mutex> lock(mu_); |
|
signal_client_ = true; |
|
++rpcs_waiting_for_client_cancel_; |
|
} |
|
while (!context->IsCancelled()) { |
|
gpr_sleep_until(gpr_time_add( |
|
gpr_now(GPR_CLOCK_REALTIME), |
|
gpr_time_from_micros(request->param().client_cancel_after_us() * |
|
grpc_test_slowdown_factor(), |
|
GPR_TIMESPAN))); |
|
} |
|
{ |
|
std::unique_lock<std::mutex> lock(mu_); |
|
--rpcs_waiting_for_client_cancel_; |
|
} |
|
return Status::CANCELLED; |
|
} else if (request->has_param() && |
|
request->param().server_cancel_after_us()) { |
|
gpr_sleep_until(gpr_time_add( |
|
gpr_now(GPR_CLOCK_REALTIME), |
|
gpr_time_from_micros(request->param().server_cancel_after_us() * |
|
grpc_test_slowdown_factor(), |
|
GPR_TIMESPAN))); |
|
return Status::CANCELLED; |
|
} else if (!request->has_param() || |
|
!request->param().skip_cancelled_check()) { |
|
EXPECT_FALSE(context->IsCancelled()); |
|
} |
|
|
|
if (request->has_param() && request->param().echo_metadata_initially()) { |
|
const std::multimap<grpc::string_ref, grpc::string_ref>& client_metadata = |
|
context->client_metadata(); |
|
for (const auto& metadatum : client_metadata) { |
|
context->AddInitialMetadata(ToString(metadatum.first), |
|
ToString(metadatum.second)); |
|
} |
|
} |
|
|
|
if (request->has_param() && request->param().echo_metadata()) { |
|
const std::multimap<grpc::string_ref, grpc::string_ref>& client_metadata = |
|
context->client_metadata(); |
|
for (const auto& metadatum : client_metadata) { |
|
context->AddTrailingMetadata(ToString(metadatum.first), |
|
ToString(metadatum.second)); |
|
} |
|
// Terminate rpc with error and debug info in trailer. |
|
if (request->param().debug_info().stack_entries_size() || |
|
!request->param().debug_info().detail().empty()) { |
|
std::string serialized_debug_info = |
|
request->param().debug_info().SerializeAsString(); |
|
context->AddTrailingMetadata(kDebugInfoTrailerKey, |
|
serialized_debug_info); |
|
return Status::CANCELLED; |
|
} |
|
} |
|
if (request->has_param() && |
|
(!request->param().expected_client_identity().empty() || |
|
request->param().check_auth_context())) { |
|
internal::CheckServerAuthContext( |
|
context, request->param().expected_transport_security_type(), |
|
request->param().expected_client_identity()); |
|
} |
|
if (request->has_param() && |
|
request->param().response_message_length() > 0) { |
|
response->set_message( |
|
std::string(request->param().response_message_length(), '\0')); |
|
} |
|
if (request->has_param() && request->param().echo_peer()) { |
|
response->mutable_param()->set_peer(context->peer()); |
|
} |
|
return Status::OK; |
|
} |
|
|
|
Status Echo1(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) { |
|
return Echo(context, request, response); |
|
} |
|
|
|
Status Echo2(ServerContext* context, const EchoRequest* request, |
|
EchoResponse* response) { |
|
return Echo(context, request, response); |
|
} |
|
|
|
Status CheckClientInitialMetadata(ServerContext* context, |
|
const SimpleRequest* /*request*/, |
|
SimpleResponse* /*response*/) { |
|
EXPECT_EQ(internal::MetadataMatchCount(context->client_metadata(), |
|
kCheckClientInitialMetadataKey, |
|
kCheckClientInitialMetadataVal), |
|
1); |
|
EXPECT_EQ(1u, |
|
context->client_metadata().count(kCheckClientInitialMetadataKey)); |
|
return Status::OK; |
|
} |
|
|
|
// Unimplemented is left unimplemented to test the returned error. |
|
|
|
Status RequestStream(ServerContext* context, |
|
ServerReader<EchoRequest>* reader, |
|
EchoResponse* response) { |
|
// If 'server_try_cancel' is set in the metadata, the RPC is cancelled by |
|
// the server by calling ServerContext::TryCancel() depending on the value: |
|
// CANCEL_BEFORE_PROCESSING: The RPC is cancelled before the server reads |
|
// any message from the client |
|
// CANCEL_DURING_PROCESSING: The RPC is cancelled while the server is |
|
// reading messages from the client |
|
// CANCEL_AFTER_PROCESSING: The RPC is cancelled after the server reads |
|
// all the messages from the client |
|
int server_try_cancel = internal::GetIntValueFromMetadata( |
|
kServerTryCancelRequest, context->client_metadata(), DO_NOT_CANCEL); |
|
|
|
EchoRequest request; |
|
response->set_message(""); |
|
|
|
if (server_try_cancel == CANCEL_BEFORE_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
std::thread* server_try_cancel_thd = nullptr; |
|
if (server_try_cancel == CANCEL_DURING_PROCESSING) { |
|
server_try_cancel_thd = |
|
new std::thread([context] { internal::ServerTryCancel(context); }); |
|
} |
|
|
|
int num_msgs_read = 0; |
|
while (reader->Read(&request)) { |
|
response->mutable_message()->append(request.message()); |
|
} |
|
LOG(INFO) << "Read: " << num_msgs_read << " messages"; |
|
|
|
if (server_try_cancel_thd != nullptr) { |
|
server_try_cancel_thd->join(); |
|
delete server_try_cancel_thd; |
|
return Status::CANCELLED; |
|
} |
|
|
|
if (server_try_cancel == CANCEL_AFTER_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
return Status::OK; |
|
} |
|
|
|
// Return 'kNumResponseStreamMsgs' messages. |
|
// TODO(yangg) make it generic by adding a parameter into EchoRequest |
|
Status ResponseStream(ServerContext* context, const EchoRequest* request, |
|
ServerWriter<EchoResponse>* writer) { |
|
// If server_try_cancel is set in the metadata, the RPC is cancelled by the |
|
// server by calling ServerContext::TryCancel() depending on the value: |
|
// CANCEL_BEFORE_PROCESSING: The RPC is cancelled before the server writes |
|
// any messages to the client |
|
// CANCEL_DURING_PROCESSING: The RPC is cancelled while the server is |
|
// writing messages to the client |
|
// CANCEL_AFTER_PROCESSING: The RPC is cancelled after the server writes |
|
// all the messages to the client |
|
int server_try_cancel = internal::GetIntValueFromMetadata( |
|
kServerTryCancelRequest, context->client_metadata(), DO_NOT_CANCEL); |
|
|
|
int server_coalescing_api = internal::GetIntValueFromMetadata( |
|
kServerUseCoalescingApi, context->client_metadata(), 0); |
|
|
|
int server_responses_to_send = internal::GetIntValueFromMetadata( |
|
kServerResponseStreamsToSend, context->client_metadata(), |
|
kServerDefaultResponseStreamsToSend); |
|
|
|
if (server_try_cancel == CANCEL_BEFORE_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
EchoResponse response; |
|
std::thread* server_try_cancel_thd = nullptr; |
|
if (server_try_cancel == CANCEL_DURING_PROCESSING) { |
|
server_try_cancel_thd = |
|
new std::thread([context] { internal::ServerTryCancel(context); }); |
|
} |
|
|
|
for (int i = 0; i < server_responses_to_send; i++) { |
|
response.set_message(request->message() + std::to_string(i)); |
|
if (i == server_responses_to_send - 1 && server_coalescing_api != 0) { |
|
writer->WriteLast(response, WriteOptions()); |
|
} else { |
|
writer->Write(response); |
|
} |
|
} |
|
|
|
if (server_try_cancel_thd != nullptr) { |
|
server_try_cancel_thd->join(); |
|
delete server_try_cancel_thd; |
|
return Status::CANCELLED; |
|
} |
|
|
|
if (server_try_cancel == CANCEL_AFTER_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
return Status::OK; |
|
} |
|
|
|
Status BidiStream(ServerContext* context, |
|
ServerReaderWriter<EchoResponse, EchoRequest>* stream) { |
|
// If server_try_cancel is set in the metadata, the RPC is cancelled by the |
|
// server by calling ServerContext::TryCancel() depending on the value: |
|
// CANCEL_BEFORE_PROCESSING: The RPC is cancelled before the server reads/ |
|
// writes any messages from/to the client |
|
// CANCEL_DURING_PROCESSING: The RPC is cancelled while the server is |
|
// reading/writing messages from/to the client |
|
// CANCEL_AFTER_PROCESSING: The RPC is cancelled after the server |
|
// reads/writes all messages from/to the client |
|
int server_try_cancel = internal::GetIntValueFromMetadata( |
|
kServerTryCancelRequest, context->client_metadata(), DO_NOT_CANCEL); |
|
|
|
int client_try_cancel = static_cast<bool>(internal::GetIntValueFromMetadata( |
|
kClientTryCancelRequest, context->client_metadata(), 0)); |
|
|
|
EchoRequest request; |
|
EchoResponse response; |
|
|
|
if (server_try_cancel == CANCEL_BEFORE_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
std::thread* server_try_cancel_thd = nullptr; |
|
if (server_try_cancel == CANCEL_DURING_PROCESSING) { |
|
server_try_cancel_thd = |
|
new std::thread([context] { internal::ServerTryCancel(context); }); |
|
} |
|
|
|
// kServerFinishAfterNReads suggests after how many reads, the server should |
|
// write the last message and send status (coalesced using WriteLast) |
|
int server_write_last = internal::GetIntValueFromMetadata( |
|
kServerFinishAfterNReads, context->client_metadata(), 0); |
|
|
|
int read_counts = 0; |
|
while (stream->Read(&request)) { |
|
read_counts++; |
|
LOG(INFO) << "recv msg " << request.message(); |
|
response.set_message(request.message()); |
|
if (read_counts == server_write_last) { |
|
stream->WriteLast(response, WriteOptions()); |
|
break; |
|
} else { |
|
stream->Write(response); |
|
} |
|
} |
|
|
|
if (client_try_cancel) { |
|
EXPECT_TRUE(context->IsCancelled()); |
|
} |
|
|
|
if (server_try_cancel_thd != nullptr) { |
|
server_try_cancel_thd->join(); |
|
delete server_try_cancel_thd; |
|
return Status::CANCELLED; |
|
} |
|
|
|
if (server_try_cancel == CANCEL_AFTER_PROCESSING) { |
|
internal::ServerTryCancel(context); |
|
return Status::CANCELLED; |
|
} |
|
|
|
return Status::OK; |
|
} |
|
|
|
// Unimplemented is left unimplemented to test the returned error. |
|
bool signal_client() { |
|
std::unique_lock<std::mutex> lock(mu_); |
|
return signal_client_; |
|
} |
|
int ClientWaitUntilNRpcsStarted(int desired_rpcs, |
|
absl::Duration timeout = absl::Minutes(1)) { |
|
return signaller_.ClientWaitUntilNRpcsStarted(desired_rpcs, timeout); |
|
} |
|
void SignalServerToContinue() { signaller_.SignalServerToContinue(); } |
|
void ResetSignaller() { signaller_.Reset(); } |
|
uint64_t RpcsWaitingForClientCancel() { |
|
std::unique_lock<std::mutex> lock(mu_); |
|
return rpcs_waiting_for_client_cancel_; |
|
} |
|
|
|
private: |
|
bool signal_client_; |
|
std::mutex mu_; |
|
TestServiceSignaller signaller_; |
|
std::unique_ptr<std::string> host_; |
|
uint64_t rpcs_waiting_for_client_cancel_ = 0; |
|
}; |
|
|
|
class CallbackTestServiceImpl |
|
: public grpc::testing::EchoTestService::CallbackService { |
|
public: |
|
CallbackTestServiceImpl() : signal_client_(false), host_() {} |
|
explicit CallbackTestServiceImpl(const std::string& host) |
|
: signal_client_(false), host_(new std::string(host)) {} |
|
|
|
ServerUnaryReactor* Echo(CallbackServerContext* context, |
|
const EchoRequest* request, |
|
EchoResponse* response) override; |
|
|
|
ServerUnaryReactor* CheckClientInitialMetadata(CallbackServerContext* context, |
|
const SimpleRequest*, |
|
SimpleResponse*) override; |
|
|
|
ServerReadReactor<EchoRequest>* RequestStream( |
|
CallbackServerContext* context, EchoResponse* response) override; |
|
|
|
ServerWriteReactor<EchoResponse>* ResponseStream( |
|
CallbackServerContext* context, const EchoRequest* request) override; |
|
|
|
ServerBidiReactor<EchoRequest, EchoResponse>* BidiStream( |
|
CallbackServerContext* context) override; |
|
|
|
// Unimplemented is left unimplemented to test the returned error. |
|
bool signal_client() { |
|
std::unique_lock<std::mutex> lock(mu_); |
|
return signal_client_; |
|
} |
|
int ClientWaitUntilNRpcsStarted(int desired_rpcs, |
|
absl::Duration timeout = absl::Minutes(1)) { |
|
return signaller_.ClientWaitUntilNRpcsStarted(desired_rpcs, timeout); |
|
} |
|
void SignalServerToContinue() { signaller_.SignalServerToContinue(); } |
|
void ResetSignaller() { signaller_.Reset(); } |
|
|
|
private: |
|
bool signal_client_; |
|
std::mutex mu_; |
|
TestServiceSignaller signaller_; |
|
std::unique_ptr<std::string> host_; |
|
}; |
|
|
|
using TestServiceImpl = |
|
TestMultipleServiceImpl<grpc::testing::EchoTestService::Service>; |
|
|
|
} // namespace testing |
|
} // namespace grpc |
|
|
|
#endif // GRPC_TEST_CPP_END2END_TEST_SERVICE_IMPL_H
|
|
|