|
|
|
/*
|
|
|
|
*
|
|
|
|
* Copyright 2018 gRPC authors.
|
|
|
|
*
|
|
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
* you may not use this file except in compliance with the License.
|
|
|
|
* You may obtain a copy of the License at
|
|
|
|
*
|
|
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
|
|
*
|
|
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
|
|
* See the License for the specific language governing permissions and
|
|
|
|
* limitations under the License.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <algorithm>
|
|
|
|
#include <functional>
|
|
|
|
#include <mutex>
|
|
|
|
#include <sstream>
|
|
|
|
#include <thread>
|
|
|
|
|
|
|
|
#include <grpcpp/channel.h>
|
|
|
|
#include <grpcpp/client_context.h>
|
|
|
|
#include <grpcpp/create_channel.h>
|
|
|
|
#include <grpcpp/generic/generic_stub.h>
|
|
|
|
#include <grpcpp/impl/codegen/proto_utils.h>
|
|
|
|
#include <grpcpp/server.h>
|
|
|
|
#include <grpcpp/server_builder.h>
|
|
|
|
#include <grpcpp/server_context.h>
|
|
|
|
#include <grpcpp/support/client_callback.h>
|
|
|
|
|
|
|
|
#include "src/core/lib/iomgr/iomgr.h"
|
|
|
|
#include "src/proto/grpc/testing/echo.grpc.pb.h"
|
|
|
|
#include "test/core/util/port.h"
|
|
|
|
#include "test/core/util/test_config.h"
|
|
|
|
#include "test/cpp/end2end/interceptors_util.h"
|
|
|
|
#include "test/cpp/end2end/test_service_impl.h"
|
|
|
|
#include "test/cpp/util/byte_buffer_proto_helper.h"
|
|
|
|
#include "test/cpp/util/string_ref_helper.h"
|
|
|
|
#include "test/cpp/util/test_credentials_provider.h"
|
|
|
|
|
|
|
|
#include <gtest/gtest.h>
|
|
|
|
|
|
|
|
// MAYBE_SKIP_TEST is a macro to determine if this particular test configuration
|
|
|
|
// should be skipped based on a decision made at SetUp time. In particular, any
|
|
|
|
// callback tests can only be run if the iomgr can run in the background or if
|
|
|
|
// the transport is in-process.
|
|
|
|
#define MAYBE_SKIP_TEST \
|
|
|
|
do { \
|
|
|
|
if (do_not_test_) { \
|
|
|
|
return; \
|
|
|
|
} \
|
|
|
|
} while (0)
|
|
|
|
|
|
|
|
namespace grpc {
|
|
|
|
namespace testing {
|
|
|
|
namespace {
|
|
|
|
|
|
|
|
enum class Protocol { INPROC, TCP };
|
|
|
|
|
|
|
|
class TestScenario {
|
|
|
|
public:
|
|
|
|
TestScenario(bool serve_callback, Protocol protocol, bool intercept,
|
|
|
|
const grpc::string& creds_type)
|
|
|
|
: callback_server(serve_callback),
|
|
|
|
protocol(protocol),
|
|
|
|
use_interceptors(intercept),
|
|
|
|
credentials_type(creds_type) {}
|
|
|
|
void Log() const;
|
|
|
|
bool callback_server;
|
|
|
|
Protocol protocol;
|
|
|
|
bool use_interceptors;
|
|
|
|
const grpc::string credentials_type;
|
|
|
|
};
|
|
|
|
|
|
|
|
static std::ostream& operator<<(std::ostream& out,
|
|
|
|
const TestScenario& scenario) {
|
|
|
|
return out << "TestScenario{callback_server="
|
|
|
|
<< (scenario.callback_server ? "true" : "false") << ",protocol="
|
|
|
|
<< (scenario.protocol == Protocol::INPROC ? "INPROC" : "TCP")
|
|
|
|
<< ",intercept=" << (scenario.use_interceptors ? "true" : "false")
|
|
|
|
<< ",creds=" << scenario.credentials_type << "}";
|
|
|
|
}
|
|
|
|
|
|
|
|
void TestScenario::Log() const {
|
|
|
|
std::ostringstream out;
|
|
|
|
out << *this;
|
|
|
|
gpr_log(GPR_DEBUG, "%s", out.str().c_str());
|
|
|
|
}
|
|
|
|
|
|
|
|
class ClientCallbackEnd2endTest
|
|
|
|
: public ::testing::TestWithParam<TestScenario> {
|
|
|
|
protected:
|
|
|
|
ClientCallbackEnd2endTest() { GetParam().Log(); }
|
|
|
|
|
|
|
|
void SetUp() override {
|
|
|
|
ServerBuilder builder;
|
|
|
|
|
|
|
|
auto server_creds = GetCredentialsProvider()->GetServerCredentials(
|
|
|
|
GetParam().credentials_type);
|
|
|
|
// TODO(vjpai): Support testing of AuthMetadataProcessor
|
|
|
|
|
|
|
|
if (GetParam().protocol == Protocol::TCP) {
|
|
|
|
picked_port_ = grpc_pick_unused_port_or_die();
|
|
|
|
server_address_ << "localhost:" << picked_port_;
|
|
|
|
builder.AddListeningPort(server_address_.str(), server_creds);
|
|
|
|
}
|
|
|
|
if (!GetParam().callback_server) {
|
|
|
|
builder.RegisterService(&service_);
|
|
|
|
} else {
|
|
|
|
builder.RegisterService(&callback_service_);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
std::vector<
|
|
|
|
std::unique_ptr<experimental::ServerInterceptorFactoryInterface>>
|
|
|
|
creators;
|
|
|
|
// Add 20 dummy server interceptors
|
|
|
|
creators.reserve(20);
|
|
|
|
for (auto i = 0; i < 20; i++) {
|
|
|
|
creators.push_back(std::unique_ptr<DummyInterceptorFactory>(
|
|
|
|
new DummyInterceptorFactory()));
|
|
|
|
}
|
|
|
|
builder.experimental().SetInterceptorCreators(std::move(creators));
|
|
|
|
}
|
|
|
|
|
|
|
|
server_ = builder.BuildAndStart();
|
|
|
|
is_server_started_ = true;
|
|
|
|
if (GetParam().protocol == Protocol::TCP &&
|
|
|
|
!grpc_iomgr_run_in_background()) {
|
|
|
|
do_not_test_ = true;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void ResetStub() {
|
|
|
|
ChannelArguments args;
|
|
|
|
auto channel_creds = GetCredentialsProvider()->GetChannelCredentials(
|
|
|
|
GetParam().credentials_type, &args);
|
|
|
|
switch (GetParam().protocol) {
|
|
|
|
case Protocol::TCP:
|
|
|
|
if (!GetParam().use_interceptors) {
|
|
|
|
channel_ = ::grpc::CreateCustomChannel(server_address_.str(),
|
|
|
|
channel_creds, args);
|
|
|
|
} else {
|
|
|
|
channel_ = CreateCustomChannelWithInterceptors(
|
|
|
|
server_address_.str(), channel_creds, args,
|
|
|
|
CreateDummyClientInterceptors());
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case Protocol::INPROC:
|
|
|
|
if (!GetParam().use_interceptors) {
|
|
|
|
channel_ = server_->InProcessChannel(args);
|
|
|
|
} else {
|
|
|
|
channel_ = server_->experimental().InProcessChannelWithInterceptors(
|
|
|
|
args, CreateDummyClientInterceptors());
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
assert(false);
|
|
|
|
}
|
|
|
|
stub_ = grpc::testing::EchoTestService::NewStub(channel_);
|
|
|
|
generic_stub_.reset(new GenericStub(channel_));
|
|
|
|
DummyInterceptor::Reset();
|
|
|
|
}
|
|
|
|
|
|
|
|
void TearDown() override {
|
|
|
|
if (is_server_started_) {
|
|
|
|
server_->Shutdown();
|
|
|
|
}
|
|
|
|
if (picked_port_ > 0) {
|
|
|
|
grpc_recycle_unused_port(picked_port_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void SendRpcs(int num_rpcs, bool with_binary_metadata) {
|
|
|
|
grpc::string test_string("");
|
|
|
|
for (int i = 0; i < num_rpcs; i++) {
|
|
|
|
EchoRequest request;
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
|
|
|
|
test_string += "Hello world. ";
|
|
|
|
request.set_message(test_string);
|
|
|
|
grpc::string val;
|
|
|
|
if (with_binary_metadata) {
|
|
|
|
request.mutable_param()->set_echo_metadata(true);
|
|
|
|
char bytes[8] = {'\0', '\1', '\2', '\3',
|
|
|
|
'\4', '\5', '\6', static_cast<char>(i)};
|
|
|
|
val = grpc::string(bytes, 8);
|
|
|
|
cli_ctx.AddMetadata("custom-bin", val);
|
|
|
|
}
|
|
|
|
|
|
|
|
cli_ctx.set_compression_algorithm(GRPC_COMPRESS_GZIP);
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx, &request, &response,
|
|
|
|
[&cli_ctx, &request, &response, &done, &mu, &cv, val,
|
|
|
|
with_binary_metadata](Status s) {
|
|
|
|
GPR_ASSERT(s.ok());
|
|
|
|
|
|
|
|
EXPECT_EQ(request.message(), response.message());
|
|
|
|
if (with_binary_metadata) {
|
|
|
|
EXPECT_EQ(
|
|
|
|
1u, cli_ctx.GetServerTrailingMetadata().count("custom-bin"));
|
|
|
|
EXPECT_EQ(val, ToString(cli_ctx.GetServerTrailingMetadata()
|
|
|
|
.find("custom-bin")
|
|
|
|
->second));
|
|
|
|
}
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void SendRpcsRawReq(int num_rpcs) {
|
|
|
|
grpc::string test_string("Hello raw world.");
|
|
|
|
EchoRequest request;
|
|
|
|
request.set_message(test_string);
|
|
|
|
std::unique_ptr<ByteBuffer> send_buf = SerializeToByteBuffer(&request);
|
|
|
|
|
|
|
|
for (int i = 0; i < num_rpcs; i++) {
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx, send_buf.get(), &response,
|
|
|
|
[&request, &response, &done, &mu, &cv](Status s) {
|
|
|
|
GPR_ASSERT(s.ok());
|
|
|
|
|
|
|
|
EXPECT_EQ(request.message(), response.message());
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void SendRpcsGeneric(int num_rpcs, bool maybe_except) {
|
|
|
|
const grpc::string kMethodName("/grpc.testing.EchoTestService/Echo");
|
|
|
|
grpc::string test_string("");
|
|
|
|
for (int i = 0; i < num_rpcs; i++) {
|
|
|
|
EchoRequest request;
|
|
|
|
std::unique_ptr<ByteBuffer> send_buf;
|
|
|
|
ByteBuffer recv_buf;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
|
|
|
|
test_string += "Hello world. ";
|
|
|
|
request.set_message(test_string);
|
|
|
|
send_buf = SerializeToByteBuffer(&request);
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
generic_stub_->experimental().UnaryCall(
|
|
|
|
&cli_ctx, kMethodName, send_buf.get(), &recv_buf,
|
|
|
|
[&request, &recv_buf, &done, &mu, &cv, maybe_except](Status s) {
|
|
|
|
GPR_ASSERT(s.ok());
|
|
|
|
|
|
|
|
EchoResponse response;
|
|
|
|
EXPECT_TRUE(ParseFromByteBuffer(&recv_buf, &response));
|
|
|
|
EXPECT_EQ(request.message(), response.message());
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
#if GRPC_ALLOW_EXCEPTIONS
|
|
|
|
if (maybe_except) {
|
|
|
|
throw - 1;
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
GPR_ASSERT(!maybe_except);
|
|
|
|
#endif
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void SendGenericEchoAsBidi(int num_rpcs, int reuses) {
|
|
|
|
const grpc::string kMethodName("/grpc.testing.EchoTestService/Echo");
|
|
|
|
grpc::string test_string("");
|
|
|
|
for (int i = 0; i < num_rpcs; i++) {
|
|
|
|
test_string += "Hello world. ";
|
|
|
|
class Client : public grpc::experimental::ClientBidiReactor<ByteBuffer,
|
|
|
|
ByteBuffer> {
|
|
|
|
public:
|
|
|
|
Client(ClientCallbackEnd2endTest* test, const grpc::string& method_name,
|
|
|
|
const grpc::string& test_str, int reuses)
|
|
|
|
: reuses_remaining_(reuses) {
|
|
|
|
activate_ = [this, test, method_name, test_str] {
|
|
|
|
if (reuses_remaining_ > 0) {
|
|
|
|
cli_ctx_.reset(new ClientContext);
|
|
|
|
reuses_remaining_--;
|
|
|
|
test->generic_stub_->experimental().PrepareBidiStreamingCall(
|
|
|
|
cli_ctx_.get(), method_name, this);
|
|
|
|
request_.set_message(test_str);
|
|
|
|
send_buf_ = SerializeToByteBuffer(&request_);
|
|
|
|
StartWrite(send_buf_.get());
|
|
|
|
StartRead(&recv_buf_);
|
|
|
|
StartCall();
|
|
|
|
} else {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
};
|
|
|
|
activate_();
|
|
|
|
}
|
|
|
|
void OnWriteDone(bool ok) override { StartWritesDone(); }
|
|
|
|
void OnReadDone(bool ok) override {
|
|
|
|
EchoResponse response;
|
|
|
|
EXPECT_TRUE(ParseFromByteBuffer(&recv_buf_, &response));
|
|
|
|
EXPECT_EQ(request_.message(), response.message());
|
|
|
|
};
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
activate_();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
EchoRequest request_;
|
|
|
|
std::unique_ptr<ByteBuffer> send_buf_;
|
|
|
|
ByteBuffer recv_buf_;
|
|
|
|
std::unique_ptr<ClientContext> cli_ctx_;
|
|
|
|
int reuses_remaining_;
|
|
|
|
std::function<void()> activate_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
} rpc{this, kMethodName, test_string, reuses};
|
|
|
|
|
|
|
|
rpc.Await();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
bool do_not_test_{false};
|
|
|
|
bool is_server_started_{false};
|
|
|
|
int picked_port_{0};
|
|
|
|
std::shared_ptr<Channel> channel_;
|
|
|
|
std::unique_ptr<grpc::testing::EchoTestService::Stub> stub_;
|
|
|
|
std::unique_ptr<grpc::GenericStub> generic_stub_;
|
|
|
|
TestServiceImpl service_;
|
|
|
|
CallbackTestServiceImpl callback_service_;
|
|
|
|
std::unique_ptr<Server> server_;
|
|
|
|
std::ostringstream server_address_;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SimpleRpc) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcs(1, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLockNested) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
std::mutex mu1, mu2, mu3;
|
|
|
|
std::condition_variable cv1, cv2, cv3;
|
|
|
|
bool done1 = false;
|
|
|
|
EchoRequest request1, request2, request3;
|
|
|
|
request1.set_message("Hello locked world1.");
|
|
|
|
EchoResponse response1;
|
|
|
|
ClientContext cli_ctx1;
|
|
|
|
{
|
|
|
|
std::unique_lock<std::mutex> l(mu1);
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx1, &request1, &response1,
|
|
|
|
[&mu1, this, &cv1, &done1, &request1, &response1](Status s1) {
|
|
|
|
std::unique_lock<std::mutex> l1(mu1);
|
|
|
|
EXPECT_TRUE(s1.ok());
|
|
|
|
EXPECT_EQ(request1.message(), response1.message());
|
|
|
|
// start the second level of nesting
|
|
|
|
std::mutex mu2;
|
|
|
|
bool done2 = false;
|
|
|
|
std::condition_variable cv2;
|
|
|
|
EchoRequest request2;
|
|
|
|
request2.set_message("Hello locked world2.");
|
|
|
|
EchoResponse response2;
|
|
|
|
ClientContext cli_ctx2;
|
|
|
|
std::unique_lock<std::mutex> l2(mu2);
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx2, &request2, &response2,
|
|
|
|
[&mu2, this, &cv2, &done2, &request2, &response2](Status s2) {
|
|
|
|
std::unique_lock<std::mutex> l2(mu2);
|
|
|
|
EXPECT_TRUE(s2.ok());
|
|
|
|
EXPECT_EQ(request2.message(), response2.message());
|
|
|
|
// start the third level of nesting
|
|
|
|
std::mutex mu3;
|
|
|
|
bool done3 = false;
|
|
|
|
std::condition_variable cv3;
|
|
|
|
EchoRequest request3;
|
|
|
|
request3.set_message("Hello locked world2.");
|
|
|
|
EchoResponse response3;
|
|
|
|
ClientContext cli_ctx3;
|
|
|
|
std::unique_lock<std::mutex> l3(mu3);
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx3, &request3, &response3,
|
|
|
|
[&mu3, &cv3, &done3, &request3, &response3](Status s3) {
|
|
|
|
std::lock_guard<std::mutex> l(mu3);
|
|
|
|
EXPECT_TRUE(s3.ok());
|
|
|
|
EXPECT_EQ(request3.message(), response3.message());
|
|
|
|
done3 = true;
|
|
|
|
cv3.notify_all();
|
|
|
|
});
|
|
|
|
done2 = true;
|
|
|
|
cv2.notify_all();
|
|
|
|
// Wait for inner most rpc to return.
|
|
|
|
while (!done3) {
|
|
|
|
cv3.wait(l3);
|
|
|
|
}
|
|
|
|
});
|
|
|
|
// Wait for second rpc to return.
|
|
|
|
while (!done2) {
|
|
|
|
cv2.wait(l2);
|
|
|
|
}
|
|
|
|
done1 = true;
|
|
|
|
cv1.notify_all();
|
|
|
|
});
|
|
|
|
}
|
|
|
|
|
|
|
|
std::unique_lock<std::mutex> l1(mu1);
|
|
|
|
while (!done1) {
|
|
|
|
cv1.wait(l1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SimpleRpcUnderLock) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
EchoRequest request;
|
|
|
|
request.set_message("Hello locked world.");
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
{
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&cli_ctx, &request, &response,
|
|
|
|
[&mu, &cv, &done, &request, &response](Status s) {
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(request.message(), response.message());
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
}
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialRpcs) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcs(10, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialRpcsRawReq) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcsRawReq(10);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SendClientInitialMetadata) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SimpleRequest request;
|
|
|
|
SimpleResponse response;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
|
|
|
|
cli_ctx.AddMetadata(kCheckClientInitialMetadataKey,
|
|
|
|
kCheckClientInitialMetadataVal);
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub_->experimental_async()->CheckClientInitialMetadata(
|
|
|
|
&cli_ctx, &request, &response, [&done, &mu, &cv](Status s) {
|
|
|
|
GPR_ASSERT(s.ok());
|
|
|
|
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SimpleRpcWithBinaryMetadata) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcs(1, true);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialRpcsWithVariedBinaryMetadataValue) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcs(10, true);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcs) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcsGeneric(10, false);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidi) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendGenericEchoAsBidi(10, 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SequentialGenericRpcsAsBidiWithReactorReuse) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendGenericEchoAsBidi(10, 10);
|
|
|
|
}
|
|
|
|
|
|
|
|
#if GRPC_ALLOW_EXCEPTIONS
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ExceptingRpc) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
SendRpcsGeneric(10, true);
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, MultipleRpcsWithVariedBinaryMetadataValue) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
std::vector<std::thread> threads;
|
|
|
|
threads.reserve(10);
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
|
|
threads.emplace_back([this] { SendRpcs(10, true); });
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
|
|
threads[i].join();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, MultipleRpcs) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
std::vector<std::thread> threads;
|
|
|
|
threads.reserve(10);
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
|
|
threads.emplace_back([this] { SendRpcs(10, false); });
|
|
|
|
}
|
|
|
|
for (int i = 0; i < 10; ++i) {
|
|
|
|
threads[i].join();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, CancelRpcBeforeStart) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
EchoRequest request;
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext context;
|
|
|
|
request.set_message("hello");
|
|
|
|
context.TryCancel();
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&context, &request, &response, [&response, &done, &mu, &cv](Status s) {
|
|
|
|
EXPECT_EQ("", response.message());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, RequestEchoServerCancel) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
EchoRequest request;
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext context;
|
|
|
|
request.set_message("hello");
|
|
|
|
context.AddMetadata(kServerTryCancelRequest,
|
|
|
|
grpc::to_string(CANCEL_BEFORE_PROCESSING));
|
|
|
|
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub_->experimental_async()->Echo(
|
|
|
|
&context, &request, &response, [&done, &mu, &cv](Status s) {
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
struct ClientCancelInfo {
|
|
|
|
bool cancel{false};
|
|
|
|
int ops_before_cancel;
|
|
|
|
|
|
|
|
ClientCancelInfo() : cancel{false} {}
|
|
|
|
explicit ClientCancelInfo(int ops) : cancel{true}, ops_before_cancel{ops} {}
|
|
|
|
};
|
|
|
|
|
|
|
|
class WriteClient : public grpc::experimental::ClientWriteReactor<EchoRequest> {
|
|
|
|
public:
|
|
|
|
WriteClient(grpc::testing::EchoTestService::Stub* stub,
|
|
|
|
ServerTryCancelRequestPhase server_try_cancel,
|
|
|
|
int num_msgs_to_send, ClientCancelInfo client_cancel = {})
|
|
|
|
: server_try_cancel_(server_try_cancel),
|
|
|
|
num_msgs_to_send_(num_msgs_to_send),
|
|
|
|
client_cancel_{client_cancel} {
|
|
|
|
grpc::string msg{"Hello server."};
|
|
|
|
for (int i = 0; i < num_msgs_to_send; i++) {
|
|
|
|
desired_ += msg;
|
|
|
|
}
|
|
|
|
if (server_try_cancel != DO_NOT_CANCEL) {
|
|
|
|
// Send server_try_cancel value in the client metadata
|
|
|
|
context_.AddMetadata(kServerTryCancelRequest,
|
|
|
|
grpc::to_string(server_try_cancel));
|
|
|
|
}
|
|
|
|
context_.set_initial_metadata_corked(true);
|
|
|
|
stub->experimental_async()->RequestStream(&context_, &response_, this);
|
|
|
|
StartCall();
|
|
|
|
request_.set_message(msg);
|
|
|
|
MaybeWrite();
|
|
|
|
}
|
|
|
|
void OnWriteDone(bool ok) override {
|
|
|
|
if (ok) {
|
|
|
|
num_msgs_sent_++;
|
|
|
|
MaybeWrite();
|
|
|
|
}
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
gpr_log(GPR_INFO, "Sent %d messages", num_msgs_sent_);
|
|
|
|
int num_to_send =
|
|
|
|
(client_cancel_.cancel)
|
|
|
|
? std::min(num_msgs_to_send_, client_cancel_.ops_before_cancel)
|
|
|
|
: num_msgs_to_send_;
|
|
|
|
switch (server_try_cancel_) {
|
|
|
|
case CANCEL_BEFORE_PROCESSING:
|
|
|
|
case CANCEL_DURING_PROCESSING:
|
|
|
|
// If the RPC is canceled by server before / during messages from the
|
|
|
|
// client, it means that the client most likely did not get a chance to
|
|
|
|
// send all the messages it wanted to send. i.e num_msgs_sent <=
|
|
|
|
// num_msgs_to_send
|
|
|
|
EXPECT_LE(num_msgs_sent_, num_to_send);
|
|
|
|
break;
|
|
|
|
case DO_NOT_CANCEL:
|
|
|
|
case CANCEL_AFTER_PROCESSING:
|
|
|
|
// If the RPC was not canceled or canceled after all messages were read
|
|
|
|
// by the server, the client did get a chance to send all its messages
|
|
|
|
EXPECT_EQ(num_msgs_sent_, num_to_send);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
assert(false);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if ((server_try_cancel_ == DO_NOT_CANCEL) && !client_cancel_.cancel) {
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(response_.message(), desired_);
|
|
|
|
} else {
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
}
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
void MaybeWrite() {
|
|
|
|
if (client_cancel_.cancel &&
|
|
|
|
num_msgs_sent_ == client_cancel_.ops_before_cancel) {
|
|
|
|
context_.TryCancel();
|
|
|
|
} else if (num_msgs_to_send_ > num_msgs_sent_ + 1) {
|
|
|
|
StartWrite(&request_);
|
|
|
|
} else if (num_msgs_to_send_ == num_msgs_sent_ + 1) {
|
|
|
|
StartWriteLast(&request_, WriteOptions());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext context_;
|
|
|
|
const ServerTryCancelRequestPhase server_try_cancel_;
|
|
|
|
int num_msgs_sent_{0};
|
|
|
|
const int num_msgs_to_send_;
|
|
|
|
grpc::string desired_;
|
|
|
|
const ClientCancelInfo client_cancel_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, RequestStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
WriteClient test{stub_.get(), DO_NOT_CANCEL, 3};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were not notified to cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ClientCancelsRequestStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
WriteClient test{stub_.get(), DO_NOT_CANCEL, 3, ClientCancelInfo{2}};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors got the cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel before doing reading the request
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelBeforeReads) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
WriteClient test{stub_.get(), CANCEL_BEFORE_PROCESSING, 1};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel while reading a request from the stream in parallel
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelDuringRead) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
WriteClient test{stub_.get(), CANCEL_DURING_PROCESSING, 10};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel after reading all the requests but before returning to the
|
|
|
|
// client
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, RequestStreamServerCancelAfterReads) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
WriteClient test{stub_.get(), CANCEL_AFTER_PROCESSING, 4};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, UnaryReactor) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
class UnaryClient : public grpc::experimental::ClientUnaryReactor {
|
|
|
|
public:
|
|
|
|
UnaryClient(grpc::testing::EchoTestService::Stub* stub) {
|
|
|
|
cli_ctx_.AddMetadata("key1", "val1");
|
|
|
|
cli_ctx_.AddMetadata("key2", "val2");
|
|
|
|
request_.mutable_param()->set_echo_metadata_initially(true);
|
|
|
|
request_.set_message("Hello metadata");
|
|
|
|
stub->experimental_async()->Echo(&cli_ctx_, &request_, &response_, this);
|
|
|
|
StartCall();
|
|
|
|
}
|
|
|
|
void OnReadInitialMetadataDone(bool ok) override {
|
|
|
|
EXPECT_TRUE(ok);
|
|
|
|
EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key1"));
|
|
|
|
EXPECT_EQ(
|
|
|
|
"val1",
|
|
|
|
ToString(cli_ctx_.GetServerInitialMetadata().find("key1")->second));
|
|
|
|
EXPECT_EQ(1u, cli_ctx_.GetServerInitialMetadata().count("key2"));
|
|
|
|
EXPECT_EQ(
|
|
|
|
"val2",
|
|
|
|
ToString(cli_ctx_.GetServerInitialMetadata().find("key2")->second));
|
|
|
|
initial_metadata_done_ = true;
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
EXPECT_TRUE(initial_metadata_done_);
|
|
|
|
EXPECT_EQ(0u, cli_ctx_.GetServerTrailingMetadata().size());
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(request_.message(), response_.message());
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext cli_ctx_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_{false};
|
|
|
|
bool initial_metadata_done_{false};
|
|
|
|
};
|
|
|
|
|
|
|
|
UnaryClient test{stub_.get()};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were not notified of a cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
class ReadClient : public grpc::experimental::ClientReadReactor<EchoResponse> {
|
|
|
|
public:
|
|
|
|
ReadClient(grpc::testing::EchoTestService::Stub* stub,
|
|
|
|
ServerTryCancelRequestPhase server_try_cancel,
|
|
|
|
ClientCancelInfo client_cancel = {})
|
|
|
|
: server_try_cancel_(server_try_cancel), client_cancel_{client_cancel} {
|
|
|
|
if (server_try_cancel_ != DO_NOT_CANCEL) {
|
|
|
|
// Send server_try_cancel value in the client metadata
|
|
|
|
context_.AddMetadata(kServerTryCancelRequest,
|
|
|
|
grpc::to_string(server_try_cancel));
|
|
|
|
}
|
|
|
|
request_.set_message("Hello client ");
|
|
|
|
stub->experimental_async()->ResponseStream(&context_, &request_, this);
|
|
|
|
if (client_cancel_.cancel &&
|
|
|
|
reads_complete_ == client_cancel_.ops_before_cancel) {
|
|
|
|
context_.TryCancel();
|
|
|
|
}
|
|
|
|
// Even if we cancel, read until failure because there might be responses
|
|
|
|
// pending
|
|
|
|
StartRead(&response_);
|
|
|
|
StartCall();
|
|
|
|
}
|
|
|
|
void OnReadDone(bool ok) override {
|
|
|
|
if (!ok) {
|
|
|
|
if (server_try_cancel_ == DO_NOT_CANCEL && !client_cancel_.cancel) {
|
|
|
|
EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
|
|
|
|
EXPECT_EQ(response_.message(),
|
|
|
|
request_.message() + grpc::to_string(reads_complete_));
|
|
|
|
reads_complete_++;
|
|
|
|
if (client_cancel_.cancel &&
|
|
|
|
reads_complete_ == client_cancel_.ops_before_cancel) {
|
|
|
|
context_.TryCancel();
|
|
|
|
}
|
|
|
|
// Even if we cancel, read until failure because there might be responses
|
|
|
|
// pending
|
|
|
|
StartRead(&response_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
gpr_log(GPR_INFO, "Read %d messages", reads_complete_);
|
|
|
|
switch (server_try_cancel_) {
|
|
|
|
case DO_NOT_CANCEL:
|
|
|
|
if (!client_cancel_.cancel || client_cancel_.ops_before_cancel >
|
|
|
|
kServerDefaultResponseStreamsToSend) {
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(reads_complete_, kServerDefaultResponseStreamsToSend);
|
|
|
|
} else {
|
|
|
|
EXPECT_GE(reads_complete_, client_cancel_.ops_before_cancel);
|
|
|
|
EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
|
|
|
|
// Status might be ok or cancelled depending on whether server
|
|
|
|
// sent status before client cancel went through
|
|
|
|
if (!s.ok()) {
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case CANCEL_BEFORE_PROCESSING:
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
EXPECT_EQ(reads_complete_, 0);
|
|
|
|
break;
|
|
|
|
case CANCEL_DURING_PROCESSING:
|
|
|
|
case CANCEL_AFTER_PROCESSING:
|
|
|
|
// If server canceled while writing messages, client must have read
|
|
|
|
// less than or equal to the expected number of messages. Even if the
|
|
|
|
// server canceled after writing all messages, the RPC may be canceled
|
|
|
|
// before the Client got a chance to read all the messages.
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
EXPECT_LE(reads_complete_, kServerDefaultResponseStreamsToSend);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
assert(false);
|
|
|
|
}
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext context_;
|
|
|
|
const ServerTryCancelRequestPhase server_try_cancel_;
|
|
|
|
int reads_complete_{0};
|
|
|
|
const ClientCancelInfo client_cancel_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ResponseStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
ReadClient test{stub_.get(), DO_NOT_CANCEL};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were not notified of a cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ClientCancelsResponseStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
ReadClient test{stub_.get(), DO_NOT_CANCEL, ClientCancelInfo{2}};
|
|
|
|
test.Await();
|
|
|
|
// Because cancel in this case races with server finish, we can't be sure that
|
|
|
|
// server interceptors even see cancellation
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel before sending any response messages
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelBefore) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
ReadClient test{stub_.get(), CANCEL_BEFORE_PROCESSING};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel while writing a response to the stream in parallel
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelDuring) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
ReadClient test{stub_.get(), CANCEL_DURING_PROCESSING};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel after writing all the respones to the stream but before
|
|
|
|
// returning to the client
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ResponseStreamServerCancelAfter) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
ReadClient test{stub_.get(), CANCEL_AFTER_PROCESSING};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
class BidiClient
|
|
|
|
: public grpc::experimental::ClientBidiReactor<EchoRequest, EchoResponse> {
|
|
|
|
public:
|
|
|
|
BidiClient(grpc::testing::EchoTestService::Stub* stub,
|
|
|
|
ServerTryCancelRequestPhase server_try_cancel,
|
|
|
|
int num_msgs_to_send, ClientCancelInfo client_cancel = {})
|
|
|
|
: server_try_cancel_(server_try_cancel),
|
|
|
|
msgs_to_send_{num_msgs_to_send},
|
|
|
|
client_cancel_{client_cancel} {
|
|
|
|
if (server_try_cancel_ != DO_NOT_CANCEL) {
|
|
|
|
// Send server_try_cancel value in the client metadata
|
|
|
|
context_.AddMetadata(kServerTryCancelRequest,
|
|
|
|
grpc::to_string(server_try_cancel));
|
|
|
|
}
|
|
|
|
request_.set_message("Hello fren ");
|
|
|
|
stub->experimental_async()->BidiStream(&context_, this);
|
|
|
|
MaybeWrite();
|
|
|
|
StartRead(&response_);
|
|
|
|
StartCall();
|
|
|
|
}
|
|
|
|
void OnReadDone(bool ok) override {
|
|
|
|
if (!ok) {
|
|
|
|
if (server_try_cancel_ == DO_NOT_CANCEL) {
|
|
|
|
if (!client_cancel_.cancel) {
|
|
|
|
EXPECT_EQ(reads_complete_, msgs_to_send_);
|
|
|
|
} else {
|
|
|
|
EXPECT_LE(reads_complete_, writes_complete_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
EXPECT_LE(reads_complete_, msgs_to_send_);
|
|
|
|
EXPECT_EQ(response_.message(), request_.message());
|
|
|
|
reads_complete_++;
|
|
|
|
StartRead(&response_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
void OnWriteDone(bool ok) override {
|
|
|
|
if (server_try_cancel_ == DO_NOT_CANCEL) {
|
|
|
|
EXPECT_TRUE(ok);
|
|
|
|
} else if (!ok) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
writes_complete_++;
|
|
|
|
MaybeWrite();
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
gpr_log(GPR_INFO, "Sent %d messages", writes_complete_);
|
|
|
|
gpr_log(GPR_INFO, "Read %d messages", reads_complete_);
|
|
|
|
switch (server_try_cancel_) {
|
|
|
|
case DO_NOT_CANCEL:
|
|
|
|
if (!client_cancel_.cancel ||
|
|
|
|
client_cancel_.ops_before_cancel > msgs_to_send_) {
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(writes_complete_, msgs_to_send_);
|
|
|
|
EXPECT_EQ(reads_complete_, writes_complete_);
|
|
|
|
} else {
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
EXPECT_EQ(writes_complete_, client_cancel_.ops_before_cancel);
|
|
|
|
EXPECT_LE(reads_complete_, writes_complete_);
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
case CANCEL_BEFORE_PROCESSING:
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
// The RPC is canceled before the server did any work or returned any
|
|
|
|
// reads, but it's possible that some writes took place first from the
|
|
|
|
// client
|
|
|
|
EXPECT_LE(writes_complete_, msgs_to_send_);
|
|
|
|
EXPECT_EQ(reads_complete_, 0);
|
|
|
|
break;
|
|
|
|
case CANCEL_DURING_PROCESSING:
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
EXPECT_LE(writes_complete_, msgs_to_send_);
|
|
|
|
EXPECT_LE(reads_complete_, writes_complete_);
|
|
|
|
break;
|
|
|
|
case CANCEL_AFTER_PROCESSING:
|
|
|
|
EXPECT_FALSE(s.ok());
|
|
|
|
EXPECT_EQ(grpc::StatusCode::CANCELLED, s.error_code());
|
|
|
|
EXPECT_EQ(writes_complete_, msgs_to_send_);
|
|
|
|
// The Server canceled after reading the last message and after writing
|
|
|
|
// the message to the client. However, the RPC cancellation might have
|
|
|
|
// taken effect before the client actually read the response.
|
|
|
|
EXPECT_LE(reads_complete_, writes_complete_);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
assert(false);
|
|
|
|
}
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
void MaybeWrite() {
|
|
|
|
if (client_cancel_.cancel &&
|
|
|
|
writes_complete_ == client_cancel_.ops_before_cancel) {
|
|
|
|
context_.TryCancel();
|
|
|
|
} else if (writes_complete_ == msgs_to_send_) {
|
|
|
|
StartWritesDone();
|
|
|
|
} else {
|
|
|
|
StartWrite(&request_);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext context_;
|
|
|
|
const ServerTryCancelRequestPhase server_try_cancel_;
|
|
|
|
int reads_complete_{0};
|
|
|
|
int writes_complete_{0};
|
|
|
|
const int msgs_to_send_;
|
|
|
|
const ClientCancelInfo client_cancel_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
};
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, BidiStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
BidiClient test{stub_.get(), DO_NOT_CANCEL,
|
|
|
|
kServerDefaultResponseStreamsToSend};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were not notified of a cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(0, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, ClientCancelsBidiStream) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
BidiClient test{stub_.get(), DO_NOT_CANCEL,
|
|
|
|
kServerDefaultResponseStreamsToSend, ClientCancelInfo{2}};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified of a cancel
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel before reading/writing any requests/responses on the stream
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelBefore) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
BidiClient test{stub_.get(), CANCEL_BEFORE_PROCESSING, 2};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel while reading/writing requests/responses on the stream in
|
|
|
|
// parallel
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelDuring) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
BidiClient test{stub_.get(), CANCEL_DURING_PROCESSING, 10};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// Server to cancel after reading/writing all requests/responses on the stream
|
|
|
|
// but before returning to the client
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, BidiStreamServerCancelAfter) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
BidiClient test{stub_.get(), CANCEL_AFTER_PROCESSING, 5};
|
|
|
|
test.Await();
|
|
|
|
// Make sure that the server interceptors were notified
|
|
|
|
if (GetParam().use_interceptors) {
|
|
|
|
EXPECT_EQ(20, DummyInterceptor::GetNumTimesCancel());
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, SimultaneousReadAndWritesDone) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
class Client : public grpc::experimental::ClientBidiReactor<EchoRequest,
|
|
|
|
EchoResponse> {
|
|
|
|
public:
|
|
|
|
Client(grpc::testing::EchoTestService::Stub* stub) {
|
|
|
|
request_.set_message("Hello bidi ");
|
|
|
|
stub->experimental_async()->BidiStream(&context_, this);
|
|
|
|
StartWrite(&request_);
|
|
|
|
StartCall();
|
|
|
|
}
|
|
|
|
void OnReadDone(bool ok) override {
|
|
|
|
EXPECT_TRUE(ok);
|
|
|
|
EXPECT_EQ(response_.message(), request_.message());
|
|
|
|
}
|
|
|
|
void OnWriteDone(bool ok) override {
|
|
|
|
EXPECT_TRUE(ok);
|
|
|
|
// Now send out the simultaneous Read and WritesDone
|
|
|
|
StartWritesDone();
|
|
|
|
StartRead(&response_);
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
EXPECT_TRUE(s.ok());
|
|
|
|
EXPECT_EQ(response_.message(), request_.message());
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
cv_.notify_one();
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext context_;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
} test{stub_.get()};
|
|
|
|
|
|
|
|
test.Await();
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest, UnimplementedRpc) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ChannelArguments args;
|
|
|
|
const auto& channel_creds = GetCredentialsProvider()->GetChannelCredentials(
|
|
|
|
GetParam().credentials_type, &args);
|
|
|
|
std::shared_ptr<Channel> channel =
|
|
|
|
(GetParam().protocol == Protocol::TCP)
|
|
|
|
? ::grpc::CreateCustomChannel(server_address_.str(), channel_creds,
|
|
|
|
args)
|
|
|
|
: server_->InProcessChannel(args);
|
|
|
|
std::unique_ptr<grpc::testing::UnimplementedEchoService::Stub> stub;
|
|
|
|
stub = grpc::testing::UnimplementedEchoService::NewStub(channel);
|
|
|
|
EchoRequest request;
|
|
|
|
EchoResponse response;
|
|
|
|
ClientContext cli_ctx;
|
|
|
|
request.set_message("Hello world.");
|
|
|
|
std::mutex mu;
|
|
|
|
std::condition_variable cv;
|
|
|
|
bool done = false;
|
|
|
|
stub->experimental_async()->Unimplemented(
|
|
|
|
&cli_ctx, &request, &response, [&done, &mu, &cv](Status s) {
|
|
|
|
EXPECT_EQ(StatusCode::UNIMPLEMENTED, s.error_code());
|
|
|
|
EXPECT_EQ("", s.error_message());
|
|
|
|
|
|
|
|
std::lock_guard<std::mutex> l(mu);
|
|
|
|
done = true;
|
|
|
|
cv.notify_one();
|
|
|
|
});
|
|
|
|
std::unique_lock<std::mutex> l(mu);
|
|
|
|
while (!done) {
|
|
|
|
cv.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
TEST_P(ClientCallbackEnd2endTest,
|
|
|
|
ResponseStreamExtraReactionFlowReadsUntilDone) {
|
|
|
|
MAYBE_SKIP_TEST;
|
|
|
|
ResetStub();
|
|
|
|
class ReadAllIncomingDataClient
|
|
|
|
: public grpc::experimental::ClientReadReactor<EchoResponse> {
|
|
|
|
public:
|
|
|
|
ReadAllIncomingDataClient(grpc::testing::EchoTestService::Stub* stub) {
|
|
|
|
request_.set_message("Hello client ");
|
|
|
|
stub->experimental_async()->ResponseStream(&context_, &request_, this);
|
|
|
|
}
|
|
|
|
bool WaitForReadDone() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!read_done_) {
|
|
|
|
read_cv_.wait(l);
|
|
|
|
}
|
|
|
|
read_done_ = false;
|
|
|
|
return read_ok_;
|
|
|
|
}
|
|
|
|
void Await() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
while (!done_) {
|
|
|
|
done_cv_.wait(l);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
const Status& status() {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
return status_;
|
|
|
|
}
|
|
|
|
|
|
|
|
private:
|
|
|
|
void OnReadDone(bool ok) override {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
read_ok_ = ok;
|
|
|
|
read_done_ = true;
|
|
|
|
read_cv_.notify_one();
|
|
|
|
}
|
|
|
|
void OnDone(const Status& s) override {
|
|
|
|
std::unique_lock<std::mutex> l(mu_);
|
|
|
|
done_ = true;
|
|
|
|
status_ = s;
|
|
|
|
done_cv_.notify_one();
|
|
|
|
}
|
|
|
|
|
|
|
|
EchoRequest request_;
|
|
|
|
EchoResponse response_;
|
|
|
|
ClientContext context_;
|
|
|
|
bool read_ok_ = false;
|
|
|
|
bool read_done_ = false;
|
|
|
|
std::mutex mu_;
|
|
|
|
std::condition_variable read_cv_;
|
|
|
|
std::condition_variable done_cv_;
|
|
|
|
bool done_ = false;
|
|
|
|
Status status_;
|
|
|
|
} client{stub_.get()};
|
|
|
|
|
|
|
|
int reads_complete = 0;
|
|
|
|
client.AddHold();
|
|
|
|
client.StartCall();
|
|
|
|
|
|
|
|
EchoResponse response;
|
|
|
|
bool read_ok = true;
|
|
|
|
while (read_ok) {
|
|
|
|
client.StartRead(&response);
|
|
|
|
read_ok = client.WaitForReadDone();
|
|
|
|
if (read_ok) {
|
|
|
|
++reads_complete;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
client.RemoveHold();
|
|
|
|
client.Await();
|
|
|
|
|
|
|
|
EXPECT_EQ(kServerDefaultResponseStreamsToSend, reads_complete);
|
|
|
|
EXPECT_EQ(client.status().error_code(), grpc::StatusCode::OK);
|
|
|
|
}
|
|
|
|
|
|
|
|
std::vector<TestScenario> CreateTestScenarios(bool test_insecure) {
|
|
|
|
std::vector<TestScenario> scenarios;
|
|
|
|
std::vector<grpc::string> credentials_types{
|
|
|
|
GetCredentialsProvider()->GetSecureCredentialsTypeList()};
|
|
|
|
auto insec_ok = [] {
|
|
|
|
// Only allow insecure credentials type when it is registered with the
|
|
|
|
// provider. User may create providers that do not have insecure.
|
|
|
|
return GetCredentialsProvider()->GetChannelCredentials(
|
|
|
|
kInsecureCredentialsType, nullptr) != nullptr;
|
|
|
|
};
|
|
|
|
if (test_insecure && insec_ok()) {
|
|
|
|
credentials_types.push_back(kInsecureCredentialsType);
|
|
|
|
}
|
|
|
|
GPR_ASSERT(!credentials_types.empty());
|
|
|
|
|
|
|
|
bool barr[]{false, true};
|
|
|
|
Protocol parr[]{Protocol::INPROC, Protocol::TCP};
|
|
|
|
for (Protocol p : parr) {
|
|
|
|
for (const auto& cred : credentials_types) {
|
|
|
|
// TODO(vjpai): Test inproc with secure credentials when feasible
|
|
|
|
if (p == Protocol::INPROC &&
|
|
|
|
(cred != kInsecureCredentialsType || !insec_ok())) {
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
for (bool callback_server : barr) {
|
|
|
|
for (bool use_interceptors : barr) {
|
|
|
|
scenarios.emplace_back(callback_server, p, use_interceptors, cred);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return scenarios;
|
|
|
|
}
|
|
|
|
|
|
|
|
INSTANTIATE_TEST_CASE_P(ClientCallbackEnd2endTest, ClientCallbackEnd2endTest,
|
|
|
|
::testing::ValuesIn(CreateTestScenarios(true)));
|
|
|
|
|
|
|
|
} // namespace
|
|
|
|
} // namespace testing
|
|
|
|
} // namespace grpc
|
|
|
|
|
|
|
|
int main(int argc, char** argv) {
|
|
|
|
grpc::testing::TestEnvironment env(argc, argv);
|
|
|
|
::testing::InitGoogleTest(&argc, argv);
|
|
|
|
return RUN_ALL_TESTS();
|
|
|
|
}
|