/* * * Copyright 2017 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */ #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include #include "src/core/ext/filters/client_channel/backup_poller.h" #include "src/core/ext/filters/client_channel/parse_address.h" #include "src/core/ext/filters/client_channel/resolver/fake/fake_resolver.h" #include "src/core/ext/filters/client_channel/server_address.h" #include "src/core/lib/gpr/env.h" #include "src/core/lib/gprpp/map.h" #include "src/core/lib/gprpp/ref_counted_ptr.h" #include "src/core/lib/gprpp/sync.h" #include "src/core/lib/iomgr/sockaddr.h" #include "src/core/lib/security/credentials/fake/fake_credentials.h" #include "src/cpp/client/secure_credentials.h" #include "src/cpp/server/secure_server_credentials.h" #include "test/core/util/port.h" #include "test/core/util/test_config.h" #include "test/cpp/end2end/test_service_impl.h" #include "src/proto/grpc/lb/v2/eds_for_test.grpc.pb.h" #include "src/proto/grpc/lb/v2/lrs_for_test.grpc.pb.h" #include "src/proto/grpc/testing/echo.grpc.pb.h" #include #include // TODO(dgq): Other scenarios in need of testing: // - Send a serverlist with faulty ip:port addresses (port > 2^16, etc). // - Test reception of invalid serverlist // - Test against a non-LB server. // - Random LB server closing the stream unexpectedly. // // Findings from end to end testing to be covered here: // - Handling of LB servers restart, including reconnection after backing-off // retries. // - Destruction of load balanced channel (and therefore of xds instance) // while: // 1) the internal LB call is still active. This should work by virtue // of the weak reference the LB call holds. The call should be terminated as // part of the xds shutdown process. // 2) the retry timer is active. Again, the weak reference it holds should // prevent a premature call to \a glb_destroy. namespace grpc { namespace testing { namespace { using std::chrono::system_clock; using ::envoy::api::v2::ClusterLoadAssignment; using ::envoy::api::v2::DiscoveryRequest; using ::envoy::api::v2::DiscoveryResponse; using ::envoy::api::v2::EndpointDiscoveryService; using ::envoy::service::load_stats::v2::ClusterStats; using ::envoy::service::load_stats::v2::LoadReportingService; using ::envoy::service::load_stats::v2::LoadStatsRequest; using ::envoy::service::load_stats::v2::LoadStatsResponse; using ::envoy::service::load_stats::v2::UpstreamLocalityStats; constexpr char kEdsTypeUrl[] = "type.googleapis.com/envoy.api.v2.ClusterLoadAssignment"; constexpr char kDefaultLocalityRegion[] = "xds_default_locality_region"; constexpr char kDefaultLocalityZone[] = "xds_default_locality_zone"; constexpr char kDefaultLocalitySubzone[] = "xds_default_locality_subzone"; template class CountedService : public ServiceType { public: size_t request_count() { grpc_core::MutexLock lock(&mu_); return request_count_; } size_t response_count() { grpc_core::MutexLock lock(&mu_); return response_count_; } void IncreaseResponseCount() { grpc_core::MutexLock lock(&mu_); ++response_count_; } void IncreaseRequestCount() { grpc_core::MutexLock lock(&mu_); ++request_count_; } void ResetCounters() { grpc_core::MutexLock lock(&mu_); request_count_ = 0; response_count_ = 0; } protected: grpc_core::Mutex mu_; private: size_t request_count_ = 0; size_t response_count_ = 0; }; using BackendService = CountedService; using EdsService = CountedService; using LrsService = CountedService; const char g_kCallCredsMdKey[] = "Balancer should not ..."; const char g_kCallCredsMdValue[] = "... receive me"; class BackendServiceImpl : public BackendService { public: BackendServiceImpl() {} Status Echo(ServerContext* context, const EchoRequest* request, EchoResponse* response) override { // Backend should receive the call credentials metadata. auto call_credentials_entry = context->client_metadata().find(g_kCallCredsMdKey); EXPECT_NE(call_credentials_entry, context->client_metadata().end()); if (call_credentials_entry != context->client_metadata().end()) { EXPECT_EQ(call_credentials_entry->second, g_kCallCredsMdValue); } IncreaseRequestCount(); const auto status = TestServiceImpl::Echo(context, request, response); IncreaseResponseCount(); AddClient(context->peer()); return status; } void Start() {} void Shutdown() {} std::set clients() { grpc_core::MutexLock lock(&clients_mu_); return clients_; } private: void AddClient(const grpc::string& client) { grpc_core::MutexLock lock(&clients_mu_); clients_.insert(client); } grpc_core::Mutex mu_; grpc_core::Mutex clients_mu_; std::set clients_; }; class ClientStats { public: struct LocalityStats { // Converts from proto message class. LocalityStats(const UpstreamLocalityStats& upstream_locality_stats) : total_successful_requests( upstream_locality_stats.total_successful_requests()), total_requests_in_progress( upstream_locality_stats.total_requests_in_progress()), total_error_requests(upstream_locality_stats.total_error_requests()), total_issued_requests( upstream_locality_stats.total_issued_requests()) {} uint64_t total_successful_requests; uint64_t total_requests_in_progress; uint64_t total_error_requests; uint64_t total_issued_requests; }; // Converts from proto message class. ClientStats(const ClusterStats& cluster_stats) : total_dropped_requests_(cluster_stats.total_dropped_requests()) { for (const auto& input_locality_stats : cluster_stats.upstream_locality_stats()) { locality_stats_.emplace(input_locality_stats.locality().sub_zone(), LocalityStats(input_locality_stats)); } } uint64_t total_successful_requests() const { uint64_t sum = 0; for (auto& p : locality_stats_) { sum += p.second.total_successful_requests; } return sum; } uint64_t total_requests_in_progress() const { uint64_t sum = 0; for (auto& p : locality_stats_) { sum += p.second.total_requests_in_progress; } return sum; } uint64_t total_error_requests() const { uint64_t sum = 0; for (auto& p : locality_stats_) { sum += p.second.total_error_requests; } return sum; } uint64_t total_issued_requests() const { uint64_t sum = 0; for (auto& p : locality_stats_) { sum += p.second.total_issued_requests; } return sum; } uint64_t total_dropped_requests() const { return total_dropped_requests_; } private: std::map locality_stats_; uint64_t total_dropped_requests_; }; class EdsServiceImpl : public EdsService { public: using Stream = ServerReaderWriter; using ResponseDelayPair = std::pair; Status StreamEndpoints(ServerContext* context, Stream* stream) override { gpr_log(GPR_INFO, "LB[%p]: EDS StreamEndpoints starts", this); [&]() { { grpc_core::MutexLock lock(&eds_mu_); if (eds_done_) return; } // Balancer shouldn't receive the call credentials metadata. EXPECT_EQ(context->client_metadata().find(g_kCallCredsMdKey), context->client_metadata().end()); // Read request. DiscoveryRequest request; if (!stream->Read(&request)) return; IncreaseRequestCount(); gpr_log(GPR_INFO, "LB[%p]: received initial message '%s'", this, request.DebugString().c_str()); // Send response. std::vector responses_and_delays; { grpc_core::MutexLock lock(&eds_mu_); responses_and_delays = responses_and_delays_; } for (const auto& response_and_delay : responses_and_delays) { SendResponse(stream, response_and_delay.first, response_and_delay.second); } // Wait until notified done. grpc_core::MutexLock lock(&eds_mu_); eds_cond_.WaitUntil(&eds_mu_, [this] { return eds_done_; }); }(); gpr_log(GPR_INFO, "LB[%p]: EDS StreamEndpoints done", this); return Status::OK; } void add_response(const DiscoveryResponse& response, int send_after_ms) { grpc_core::MutexLock lock(&eds_mu_); responses_and_delays_.push_back(std::make_pair(response, send_after_ms)); } void Start() { grpc_core::MutexLock lock(&eds_mu_); eds_done_ = false; responses_and_delays_.clear(); } void Shutdown() { { grpc_core::MutexLock lock(&eds_mu_); NotifyDoneWithEdsCallLocked(); responses_and_delays_.clear(); } gpr_log(GPR_INFO, "LB[%p]: shut down", this); } static DiscoveryResponse BuildResponseForBackends( const std::vector>& backend_ports) { ClusterLoadAssignment assignment; assignment.set_cluster_name("service name"); for (size_t i = 0; i < backend_ports.size(); ++i) { auto* endpoints = assignment.add_endpoints(); endpoints->mutable_load_balancing_weight()->set_value(3); endpoints->set_priority(0); endpoints->mutable_locality()->set_region(kDefaultLocalityRegion); endpoints->mutable_locality()->set_zone(kDefaultLocalityZone); std::ostringstream sub_zone; sub_zone << kDefaultLocalitySubzone << '_' << i; endpoints->mutable_locality()->set_sub_zone(sub_zone.str()); for (const int& backend_port : backend_ports[i]) { auto* lb_endpoints = endpoints->add_lb_endpoints(); auto* endpoint = lb_endpoints->mutable_endpoint(); auto* address = endpoint->mutable_address(); auto* socket_address = address->mutable_socket_address(); socket_address->set_address("127.0.0.1"); socket_address->set_port_value(backend_port); } } DiscoveryResponse response; response.set_type_url(kEdsTypeUrl); response.add_resources()->PackFrom(assignment); return response; } void NotifyDoneWithEdsCall() { grpc_core::MutexLock lock(&eds_mu_); NotifyDoneWithEdsCallLocked(); } void NotifyDoneWithEdsCallLocked() { if (!eds_done_) { eds_done_ = true; eds_cond_.Broadcast(); } } private: void SendResponse(Stream* stream, const DiscoveryResponse& response, int delay_ms) { gpr_log(GPR_INFO, "LB[%p]: sleeping for %d ms...", this, delay_ms); if (delay_ms > 0) { gpr_sleep_until(grpc_timeout_milliseconds_to_deadline(delay_ms)); } gpr_log(GPR_INFO, "LB[%p]: Woke up! Sending response '%s'", this, response.DebugString().c_str()); IncreaseResponseCount(); stream->Write(response); } grpc_core::CondVar eds_cond_; // Protect the members below. grpc_core::Mutex eds_mu_; bool eds_done_ = false; std::vector responses_and_delays_; }; class LrsServiceImpl : public LrsService { public: using Stream = ServerReaderWriter; explicit LrsServiceImpl(int client_load_reporting_interval_seconds) : client_load_reporting_interval_seconds_( client_load_reporting_interval_seconds) {} Status StreamLoadStats(ServerContext* context, Stream* stream) override { gpr_log(GPR_INFO, "LB[%p]: LRS StreamLoadStats starts", this); // Read request. LoadStatsRequest request; if (stream->Read(&request)) { if (client_load_reporting_interval_seconds_ > 0) { IncreaseRequestCount(); // Send response. LoadStatsResponse response; auto server_name = request.cluster_stats()[0].cluster_name(); GPR_ASSERT(server_name != ""); response.add_clusters(server_name); response.mutable_load_reporting_interval()->set_seconds( client_load_reporting_interval_seconds_); stream->Write(response); IncreaseResponseCount(); // Wait for report. request.Clear(); if (stream->Read(&request)) { gpr_log(GPR_INFO, "LB[%p]: received client load report message '%s'", this, request.DebugString().c_str()); GPR_ASSERT(request.cluster_stats().size() == 1); const ClusterStats& cluster_stats = request.cluster_stats()[0]; // We need to acquire the lock here in order to prevent the notify_one // below from firing before its corresponding wait is executed. grpc_core::MutexLock lock(&load_report_mu_); GPR_ASSERT(client_stats_ == nullptr); client_stats_.reset(new ClientStats(cluster_stats)); load_report_ready_ = true; load_report_cond_.Signal(); } } // Wait until notified done. grpc_core::MutexLock lock(&lrs_mu_); lrs_cv_.WaitUntil(&lrs_mu_, [this] { return lrs_done; }); } gpr_log(GPR_INFO, "LB[%p]: LRS done", this); return Status::OK; } void Start() { lrs_done = false; load_report_ready_ = false; client_stats_.reset(); } void Shutdown() { { grpc_core::MutexLock lock(&lrs_mu_); NotifyDoneWithLrsCallLocked(); } gpr_log(GPR_INFO, "LB[%p]: shut down", this); } ClientStats* WaitForLoadReport() { grpc_core::MutexLock lock(&load_report_mu_); load_report_cond_.WaitUntil(&load_report_mu_, [this] { return load_report_ready_; }); load_report_ready_ = false; return client_stats_.get(); } void NotifyDoneWithLrsCall() { grpc_core::MutexLock lock(&lrs_mu_); NotifyDoneWithLrsCallLocked(); } void NotifyDoneWithLrsCallLocked() { if (!lrs_done) { lrs_done = true; lrs_cv_.Broadcast(); } } private: const int client_load_reporting_interval_seconds_; grpc_core::CondVar lrs_cv_; // Protect lrs_done. grpc_core::Mutex lrs_mu_; bool lrs_done = false; grpc_core::CondVar load_report_cond_; // Protect the members below. grpc_core::Mutex load_report_mu_; std::unique_ptr client_stats_; bool load_report_ready_ = false; }; class XdsEnd2endTest : public ::testing::Test { protected: XdsEnd2endTest(size_t num_backends, size_t num_balancers, int client_load_reporting_interval_seconds) : server_host_("localhost"), num_backends_(num_backends), num_balancers_(num_balancers), client_load_reporting_interval_seconds_( client_load_reporting_interval_seconds) {} static void SetUpTestCase() { // Make the backup poller poll very frequently in order to pick up // updates from all the subchannels's FDs. GPR_GLOBAL_CONFIG_SET(grpc_client_channel_backup_poll_interval_ms, 1); grpc_init(); } static void TearDownTestCase() { grpc_shutdown(); } void SetUp() override { response_generator_ = grpc_core::MakeRefCounted(); lb_channel_response_generator_ = grpc_core::MakeRefCounted(); // Start the backends. for (size_t i = 0; i < num_backends_; ++i) { backends_.emplace_back(new BackendServerThread); backends_.back()->Start(server_host_); } // Start the load balancers. for (size_t i = 0; i < num_balancers_; ++i) { balancers_.emplace_back( new BalancerServerThread(client_load_reporting_interval_seconds_)); balancers_.back()->Start(server_host_); } ResetStub(); } void TearDown() override { ShutdownAllBackends(); for (auto& balancer : balancers_) balancer->Shutdown(); } void StartAllBackends() { for (auto& backend : backends_) backend->Start(server_host_); } void StartBackend(size_t index) { backends_[index]->Start(server_host_); } void ShutdownAllBackends() { for (auto& backend : backends_) backend->Shutdown(); } void ShutdownBackend(size_t index) { backends_[index]->Shutdown(); } void ResetStub(int fallback_timeout = 0, const grpc::string& expected_targets = "") { ChannelArguments args; // TODO(juanlishen): Add setter to ChannelArguments. if (fallback_timeout > 0) { args.SetInt(GRPC_ARG_XDS_FALLBACK_TIMEOUT_MS, fallback_timeout); } args.SetPointer(GRPC_ARG_FAKE_RESOLVER_RESPONSE_GENERATOR, response_generator_.get()); if (!expected_targets.empty()) { args.SetString(GRPC_ARG_FAKE_SECURITY_EXPECTED_TARGETS, expected_targets); } std::ostringstream uri; uri << "fake:///" << kApplicationTargetName_; // TODO(dgq): templatize tests to run everything using both secure and // insecure channel credentials. grpc_channel_credentials* channel_creds = grpc_fake_transport_security_credentials_create(); grpc_call_credentials* call_creds = grpc_md_only_test_credentials_create( g_kCallCredsMdKey, g_kCallCredsMdValue, false); std::shared_ptr creds( new SecureChannelCredentials(grpc_composite_channel_credentials_create( channel_creds, call_creds, nullptr))); call_creds->Unref(); channel_creds->Unref(); channel_ = ::grpc::CreateCustomChannel(uri.str(), creds, args); stub_ = grpc::testing::EchoTestService::NewStub(channel_); } void ResetBackendCounters() { for (auto& backend : backends_) backend->backend_service()->ResetCounters(); } bool SeenAllBackends(size_t start_index = 0, size_t stop_index = 0) { if (stop_index == 0) stop_index = backends_.size(); for (size_t i = start_index; i < stop_index; ++i) { if (backends_[i]->backend_service()->request_count() == 0) return false; } return true; } void SendRpcAndCount(int* num_total, int* num_ok, int* num_failure, int* num_drops) { const Status status = SendRpc(); if (status.ok()) { ++*num_ok; } else { if (status.error_message() == "Call dropped by load balancing policy") { ++*num_drops; } else { ++*num_failure; } } ++*num_total; } std::tuple WaitForAllBackends(int num_requests_multiple_of = 1, size_t start_index = 0, size_t stop_index = 0) { int num_ok = 0; int num_failure = 0; int num_drops = 0; int num_total = 0; while (!SeenAllBackends(start_index, stop_index)) { SendRpcAndCount(&num_total, &num_ok, &num_failure, &num_drops); } while (num_total % num_requests_multiple_of != 0) { SendRpcAndCount(&num_total, &num_ok, &num_failure, &num_drops); } ResetBackendCounters(); gpr_log(GPR_INFO, "Performed %d warm up requests (a multiple of %d) against the " "backends. %d succeeded, %d failed, %d dropped.", num_total, num_requests_multiple_of, num_ok, num_failure, num_drops); return std::make_tuple(num_ok, num_failure, num_drops); } void WaitForBackend(size_t backend_idx) { do { (void)SendRpc(); } while (backends_[backend_idx]->backend_service()->request_count() == 0); ResetBackendCounters(); } grpc_core::ServerAddressList CreateLbAddressesFromPortList( const std::vector& ports) { grpc_core::ServerAddressList addresses; for (int port : ports) { char* lb_uri_str; gpr_asprintf(&lb_uri_str, "ipv4:127.0.0.1:%d", port); grpc_uri* lb_uri = grpc_uri_parse(lb_uri_str, true); GPR_ASSERT(lb_uri != nullptr); grpc_resolved_address address; GPR_ASSERT(grpc_parse_uri(lb_uri, &address)); std::vector args_to_add; grpc_channel_args* args = grpc_channel_args_copy_and_add( nullptr, args_to_add.data(), args_to_add.size()); addresses.emplace_back(address.addr, address.len, args); grpc_uri_destroy(lb_uri); gpr_free(lb_uri_str); } return addresses; } void SetNextResolution(const std::vector& ports, const char* service_config_json = nullptr, grpc_core::FakeResolverResponseGenerator* lb_channel_response_generator = nullptr) { grpc_core::ExecCtx exec_ctx; grpc_core::Resolver::Result result; result.addresses = CreateLbAddressesFromPortList(ports); if (service_config_json != nullptr) { grpc_error* error = GRPC_ERROR_NONE; result.service_config = grpc_core::ServiceConfig::Create(service_config_json, &error); GRPC_ERROR_UNREF(error); } grpc_arg arg = grpc_core::FakeResolverResponseGenerator::MakeChannelArg( lb_channel_response_generator == nullptr ? lb_channel_response_generator_.get() : lb_channel_response_generator); result.args = grpc_channel_args_copy_and_add(nullptr, &arg, 1); response_generator_->SetResponse(std::move(result)); } void SetNextResolutionForLbChannelAllBalancers( const char* service_config_json = nullptr, grpc_core::FakeResolverResponseGenerator* lb_channel_response_generator = nullptr) { std::vector ports; for (size_t i = 0; i < balancers_.size(); ++i) { ports.emplace_back(balancers_[i]->port()); } SetNextResolutionForLbChannel(ports, service_config_json, lb_channel_response_generator); } void SetNextResolutionForLbChannel( const std::vector& ports, const char* service_config_json = nullptr, grpc_core::FakeResolverResponseGenerator* lb_channel_response_generator = nullptr) { grpc_core::ExecCtx exec_ctx; grpc_core::Resolver::Result result; result.addresses = CreateLbAddressesFromPortList(ports); if (service_config_json != nullptr) { grpc_error* error = GRPC_ERROR_NONE; result.service_config = grpc_core::ServiceConfig::Create(service_config_json, &error); GRPC_ERROR_UNREF(error); } if (lb_channel_response_generator == nullptr) { lb_channel_response_generator = lb_channel_response_generator_.get(); } lb_channel_response_generator->SetResponse(std::move(result)); } void SetNextReresolutionResponse(const std::vector& ports) { grpc_core::ExecCtx exec_ctx; grpc_core::Resolver::Result result; result.addresses = CreateLbAddressesFromPortList(ports); response_generator_->SetReresolutionResponse(std::move(result)); } const std::vector GetBackendPorts(size_t start_index = 0, size_t stop_index = 0) const { if (stop_index == 0) stop_index = backends_.size(); std::vector backend_ports; for (size_t i = start_index; i < stop_index; ++i) { backend_ports.push_back(backends_[i]->port()); } return backend_ports; } const std::vector> GetBackendPortsInGroups( size_t start_index = 0, size_t stop_index = 0, size_t num_group = 1) const { if (stop_index == 0) stop_index = backends_.size(); size_t group_size = (stop_index - start_index) / num_group; std::vector> backend_ports; for (size_t i = 0; i < num_group; ++i) { backend_ports.emplace_back(); size_t group_start = group_size * i + start_index; size_t group_stop = i == num_group - 1 ? stop_index : group_start + group_size; for (size_t j = group_start; j < group_stop; ++j) { backend_ports[i].push_back(backends_[j]->port()); } } return backend_ports; } void ScheduleResponseForBalancer(size_t i, const DiscoveryResponse& response, int delay_ms) { balancers_[i]->eds_service()->add_response(response, delay_ms); } Status SendRpc(EchoResponse* response = nullptr, int timeout_ms = 1000, bool wait_for_ready = false) { const bool local_response = (response == nullptr); if (local_response) response = new EchoResponse; EchoRequest request; request.set_message(kRequestMessage_); ClientContext context; context.set_deadline(grpc_timeout_milliseconds_to_deadline(timeout_ms)); if (wait_for_ready) context.set_wait_for_ready(true); Status status = stub_->Echo(&context, request, response); if (local_response) delete response; return status; } void CheckRpcSendOk(const size_t times = 1, const int timeout_ms = 1000, bool wait_for_ready = false) { for (size_t i = 0; i < times; ++i) { EchoResponse response; const Status status = SendRpc(&response, timeout_ms, wait_for_ready); EXPECT_TRUE(status.ok()) << "code=" << status.error_code() << " message=" << status.error_message(); EXPECT_EQ(response.message(), kRequestMessage_); } } void CheckRpcSendFailure() { const Status status = SendRpc(); EXPECT_FALSE(status.ok()); } class ServerThread { public: ServerThread() : port_(grpc_pick_unused_port_or_die()) {} virtual ~ServerThread(){}; void Start(const grpc::string& server_host) { gpr_log(GPR_INFO, "starting %s server on port %d", Type(), port_); GPR_ASSERT(!running_); running_ = true; StartAllServices(); grpc_core::Mutex mu; // We need to acquire the lock here in order to prevent the notify_one // by ServerThread::Serve from firing before the wait below is hit. grpc_core::MutexLock lock(&mu); grpc_core::CondVar cond; thread_.reset(new std::thread( std::bind(&ServerThread::Serve, this, server_host, &mu, &cond))); cond.Wait(&mu); gpr_log(GPR_INFO, "%s server startup complete", Type()); } void Serve(const grpc::string& server_host, grpc_core::Mutex* mu, grpc_core::CondVar* cond) { // We need to acquire the lock here in order to prevent the notify_one // below from firing before its corresponding wait is executed. grpc_core::MutexLock lock(mu); std::ostringstream server_address; server_address << server_host << ":" << port_; ServerBuilder builder; std::shared_ptr creds(new SecureServerCredentials( grpc_fake_transport_security_server_credentials_create())); builder.AddListeningPort(server_address.str(), creds); RegisterAllServices(&builder); server_ = builder.BuildAndStart(); cond->Signal(); } void Shutdown() { if (!running_) return; gpr_log(GPR_INFO, "%s about to shutdown", Type()); ShutdownAllServices(); server_->Shutdown(grpc_timeout_milliseconds_to_deadline(0)); thread_->join(); gpr_log(GPR_INFO, "%s shutdown completed", Type()); running_ = false; } int port() const { return port_; } private: virtual void RegisterAllServices(ServerBuilder* builder) = 0; virtual void StartAllServices() = 0; virtual void ShutdownAllServices() = 0; virtual const char* Type() = 0; const int port_; std::unique_ptr server_; std::unique_ptr thread_; bool running_ = false; }; class BackendServerThread : public ServerThread { public: BackendServiceImpl* backend_service() { return &backend_service_; } private: void RegisterAllServices(ServerBuilder* builder) override { builder->RegisterService(&backend_service_); } void StartAllServices() override { backend_service_.Start(); } void ShutdownAllServices() override { backend_service_.Shutdown(); } const char* Type() override { return "Backend"; } BackendServiceImpl backend_service_; }; class BalancerServerThread : public ServerThread { public: explicit BalancerServerThread(int client_load_reporting_interval = 0) : lrs_service_(client_load_reporting_interval) {} EdsServiceImpl* eds_service() { return &eds_service_; } LrsServiceImpl* lrs_service() { return &lrs_service_; } private: void RegisterAllServices(ServerBuilder* builder) override { builder->RegisterService(&eds_service_); builder->RegisterService(&lrs_service_); } void StartAllServices() override { eds_service_.Start(); lrs_service_.Start(); } void ShutdownAllServices() override { eds_service_.Shutdown(); lrs_service_.Shutdown(); } const char* Type() override { return "Balancer"; } EdsServiceImpl eds_service_; LrsServiceImpl lrs_service_; }; const grpc::string server_host_; const size_t num_backends_; const size_t num_balancers_; const int client_load_reporting_interval_seconds_; std::shared_ptr channel_; std::unique_ptr stub_; std::vector> backends_; std::vector> balancers_; grpc_core::RefCountedPtr response_generator_; grpc_core::RefCountedPtr lb_channel_response_generator_; const grpc::string kRequestMessage_ = "Live long and prosper."; const grpc::string kApplicationTargetName_ = "application_target_name"; const grpc::string kDefaultServiceConfig_ = "{\n" " \"loadBalancingConfig\":[\n" " { \"does_not_exist\":{} },\n" " { \"xds_experimental\":{ \"balancerName\": \"fake:///lb\" } }\n" " ]\n" "}"; }; class SingleBalancerTest : public XdsEnd2endTest { public: SingleBalancerTest() : XdsEnd2endTest(4, 1, 0) {} }; TEST_F(SingleBalancerTest, Vanilla) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); const size_t kNumRpcsPerAddress = 100; ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(GetBackendPortsInGroups()), 0); // Make sure that trying to connect works without a call. channel_->GetState(true /* try_to_connect */); // We need to wait for all backends to come online. WaitForAllBackends(); // Send kNumRpcsPerAddress RPCs per server. CheckRpcSendOk(kNumRpcsPerAddress * num_backends_); // Each backend should have gotten 100 requests. for (size_t i = 0; i < backends_.size(); ++i) { EXPECT_EQ(kNumRpcsPerAddress, backends_[i]->backend_service()->request_count()); } balancers_[0]->eds_service()->NotifyDoneWithEdsCall(); // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); // Check LB policy name for the channel. EXPECT_EQ("xds_experimental", channel_->GetLoadBalancingPolicyName()); } TEST_F(SingleBalancerTest, SameBackendListedMultipleTimes) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); // Same backend listed twice. std::vector ports; ports.push_back(backends_[0]->port()); ports.push_back(backends_[0]->port()); const size_t kNumRpcsPerAddress = 10; ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends({ports}), 0); // We need to wait for the backend to come online. WaitForBackend(0); // Send kNumRpcsPerAddress RPCs per server. CheckRpcSendOk(kNumRpcsPerAddress * ports.size()); // Backend should have gotten 20 requests. EXPECT_EQ(kNumRpcsPerAddress * 2, backends_[0]->backend_service()->request_count()); // And they should have come from a single client port, because of // subchannel sharing. EXPECT_EQ(1UL, backends_[0]->backend_service()->clients().size()); balancers_[0]->eds_service()->NotifyDoneWithEdsCall(); } TEST_F(SingleBalancerTest, SecureNaming) { // TODO(juanlishen): Use separate fake creds for the balancer channel. ResetStub(0, kApplicationTargetName_ + ";lb"); SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({balancers_[0]->port()}); const size_t kNumRpcsPerAddress = 100; ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(GetBackendPortsInGroups()), 0); // Make sure that trying to connect works without a call. channel_->GetState(true /* try_to_connect */); // We need to wait for all backends to come online. WaitForAllBackends(); // Send kNumRpcsPerAddress RPCs per server. CheckRpcSendOk(kNumRpcsPerAddress * num_backends_); // Each backend should have gotten 100 requests. for (size_t i = 0; i < backends_.size(); ++i) { EXPECT_EQ(kNumRpcsPerAddress, backends_[i]->backend_service()->request_count()); } // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); } TEST_F(SingleBalancerTest, SecureNamingDeathTest) { ::testing::FLAGS_gtest_death_test_style = "threadsafe"; // Make sure that we blow up (via abort() from the security connector) when // the name from the balancer doesn't match expectations. ASSERT_DEATH_IF_SUPPORTED( { ResetStub(0, kApplicationTargetName_ + ";lb"); SetNextResolution({}, "{\n" " \"loadBalancingConfig\":[\n" " { \"does_not_exist\":{} },\n" " { \"xds_experimental\":{ \"balancerName\": " "\"fake:///wrong_lb\" } }\n" " ]\n" "}"); SetNextResolutionForLbChannel({balancers_[0]->port()}); channel_->WaitForConnected(grpc_timeout_seconds_to_deadline(1)); }, ""); } TEST_F(SingleBalancerTest, InitiallyEmptyServerlist) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); const int kServerlistDelayMs = 500 * grpc_test_slowdown_factor(); const int kCallDeadlineMs = kServerlistDelayMs * 2; // First response is an empty serverlist, sent right away. ScheduleResponseForBalancer(0, EdsServiceImpl::BuildResponseForBackends({{}}), 0); // Send non-empty serverlist only after kServerlistDelayMs ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(GetBackendPortsInGroups()), kServerlistDelayMs); const auto t0 = system_clock::now(); // Client will block: LB will initially send empty serverlist. CheckRpcSendOk(1, kCallDeadlineMs, true /* wait_for_ready */); const auto ellapsed_ms = std::chrono::duration_cast( system_clock::now() - t0); // but eventually, the LB sends a serverlist update that allows the call to // proceed. The call delay must be larger than the delay in sending the // populated serverlist but under the call's deadline (which is enforced by // the call's deadline). EXPECT_GT(ellapsed_ms.count(), kServerlistDelayMs); balancers_[0]->eds_service()->NotifyDoneWithEdsCall(); // The EDS service got a single request. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); // and sent two responses. EXPECT_EQ(2U, balancers_[0]->eds_service()->response_count()); } TEST_F(SingleBalancerTest, AllServersUnreachableFailFast) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); const size_t kNumUnreachableServers = 5; std::vector ports; for (size_t i = 0; i < kNumUnreachableServers; ++i) { ports.push_back(grpc_pick_unused_port_or_die()); } ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends({ports}), 0); const Status status = SendRpc(); // The error shouldn't be DEADLINE_EXCEEDED. EXPECT_EQ(StatusCode::UNAVAILABLE, status.error_code()); balancers_[0]->eds_service()->NotifyDoneWithEdsCall(); // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); } TEST_F(SingleBalancerTest, Fallback) { const int kFallbackTimeoutMs = 200 * grpc_test_slowdown_factor(); const int kServerlistDelayMs = 500 * grpc_test_slowdown_factor(); const size_t kNumBackendsInResolution = backends_.size() / 2; ResetStub(kFallbackTimeoutMs); SetNextResolution(GetBackendPorts(0, kNumBackendsInResolution), kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); // Send non-empty serverlist only after kServerlistDelayMs. ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends( GetBackendPortsInGroups(kNumBackendsInResolution /* start_index */)), kServerlistDelayMs); // Wait until all the fallback backends are reachable. WaitForAllBackends(1 /* num_requests_multiple_of */, 0 /* start_index */, kNumBackendsInResolution /* stop_index */); gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(kNumBackendsInResolution); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // Fallback is used: each backend returned by the resolver should have // gotten one request. for (size_t i = 0; i < kNumBackendsInResolution; ++i) { EXPECT_EQ(1U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution; i < backends_.size(); ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } // Wait until the serverlist reception has been processed and all backends // in the serverlist are reachable. WaitForAllBackends(1 /* num_requests_multiple_of */, kNumBackendsInResolution /* start_index */); gpr_log(GPR_INFO, "========= BEFORE SECOND BATCH =========="); CheckRpcSendOk(backends_.size() - kNumBackendsInResolution); gpr_log(GPR_INFO, "========= DONE WITH SECOND BATCH =========="); // Serverlist is used: each backend returned by the balancer should // have gotten one request. for (size_t i = 0; i < kNumBackendsInResolution; ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution; i < backends_.size(); ++i) { EXPECT_EQ(1U, backends_[i]->backend_service()->request_count()); } // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); } TEST_F(SingleBalancerTest, FallbackUpdate) { const int kFallbackTimeoutMs = 200 * grpc_test_slowdown_factor(); const int kServerlistDelayMs = 500 * grpc_test_slowdown_factor(); const size_t kNumBackendsInResolution = backends_.size() / 3; const size_t kNumBackendsInResolutionUpdate = backends_.size() / 3; ResetStub(kFallbackTimeoutMs); SetNextResolution(GetBackendPorts(0, kNumBackendsInResolution), kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); // Send non-empty serverlist only after kServerlistDelayMs. ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(GetBackendPortsInGroups( kNumBackendsInResolution + kNumBackendsInResolutionUpdate /* start_index */)), kServerlistDelayMs); // Wait until all the fallback backends are reachable. WaitForAllBackends(1 /* num_requests_multiple_of */, 0 /* start_index */, kNumBackendsInResolution /* stop_index */); gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(kNumBackendsInResolution); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // Fallback is used: each backend returned by the resolver should have // gotten one request. for (size_t i = 0; i < kNumBackendsInResolution; ++i) { EXPECT_EQ(1U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution; i < backends_.size(); ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } SetNextResolution(GetBackendPorts(kNumBackendsInResolution, kNumBackendsInResolution + kNumBackendsInResolutionUpdate), kDefaultServiceConfig_.c_str()); // Wait until the resolution update has been processed and all the new // fallback backends are reachable. WaitForAllBackends(1 /* num_requests_multiple_of */, kNumBackendsInResolution /* start_index */, kNumBackendsInResolution + kNumBackendsInResolutionUpdate /* stop_index */); gpr_log(GPR_INFO, "========= BEFORE SECOND BATCH =========="); CheckRpcSendOk(kNumBackendsInResolutionUpdate); gpr_log(GPR_INFO, "========= DONE WITH SECOND BATCH =========="); // The resolution update is used: each backend in the resolution update should // have gotten one request. for (size_t i = 0; i < kNumBackendsInResolution; ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution; i < kNumBackendsInResolution + kNumBackendsInResolutionUpdate; ++i) { EXPECT_EQ(1U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution + kNumBackendsInResolutionUpdate; i < backends_.size(); ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } // Wait until the serverlist reception has been processed and all backends // in the serverlist are reachable. WaitForAllBackends(1 /* num_requests_multiple_of */, kNumBackendsInResolution + kNumBackendsInResolutionUpdate /* start_index */); gpr_log(GPR_INFO, "========= BEFORE THIRD BATCH =========="); CheckRpcSendOk(backends_.size() - kNumBackendsInResolution - kNumBackendsInResolutionUpdate); gpr_log(GPR_INFO, "========= DONE WITH THIRD BATCH =========="); // Serverlist is used: each backend returned by the balancer should // have gotten one request. for (size_t i = 0; i < kNumBackendsInResolution + kNumBackendsInResolutionUpdate; ++i) { EXPECT_EQ(0U, backends_[i]->backend_service()->request_count()); } for (size_t i = kNumBackendsInResolution + kNumBackendsInResolutionUpdate; i < backends_.size(); ++i) { EXPECT_EQ(1U, backends_[i]->backend_service()->request_count()); } // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); } TEST_F(SingleBalancerTest, FallbackEarlyWhenBalancerChannelFails) { const int kFallbackTimeoutMs = 10000 * grpc_test_slowdown_factor(); ResetStub(kFallbackTimeoutMs); // Return an unreachable balancer and one fallback backend. SetNextResolution({backends_[0]->port()}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({grpc_pick_unused_port_or_die()}); // Send RPC with deadline less than the fallback timeout and make sure it // succeeds. CheckRpcSendOk(/* times */ 1, /* timeout_ms */ 1000, /* wait_for_ready */ false); } TEST_F(SingleBalancerTest, FallbackEarlyWhenBalancerCallFails) { const int kFallbackTimeoutMs = 10000 * grpc_test_slowdown_factor(); ResetStub(kFallbackTimeoutMs); // Return one balancer and one fallback backend. SetNextResolution({backends_[0]->port()}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); // Balancer drops call without sending a serverlist. balancers_[0]->eds_service()->NotifyDoneWithEdsCall(); // Send RPC with deadline less than the fallback timeout and make sure it // succeeds. CheckRpcSendOk(/* times */ 1, /* timeout_ms */ 1000, /* wait_for_ready */ false); } TEST_F(SingleBalancerTest, FallbackIfResponseReceivedButChildNotReady) { const int kFallbackTimeoutMs = 500 * grpc_test_slowdown_factor(); ResetStub(kFallbackTimeoutMs); SetNextResolution({backends_[0]->port()}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); // Send a serverlist that only contains an unreachable backend before fallback // timeout. ScheduleResponseForBalancer(0, EdsServiceImpl::BuildResponseForBackends( {{grpc_pick_unused_port_or_die()}}), 0); // Because no child policy is ready before fallback timeout, we enter fallback // mode. WaitForBackend(0); } TEST_F(SingleBalancerTest, FallbackModeIsExitedWhenBalancerSaysToDropAllCalls) { // Return an unreachable balancer and one fallback backend. SetNextResolution({backends_[0]->port()}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({grpc_pick_unused_port_or_die()}); // Enter fallback mode because the LB channel fails to connect. WaitForBackend(0); // Return a new balancer that sends an empty serverlist. ScheduleResponseForBalancer(0, EdsServiceImpl::BuildResponseForBackends({{}}), 0); SetNextResolutionForLbChannelAllBalancers(); // Send RPCs until failure. gpr_timespec deadline = gpr_time_add( gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_millis(5000, GPR_TIMESPAN)); do { auto status = SendRpc(); if (!status.ok()) break; } while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0); CheckRpcSendFailure(); } TEST_F(SingleBalancerTest, FallbackModeIsExitedAfterChildRready) { // Return an unreachable balancer and one fallback backend. SetNextResolution({backends_[0]->port()}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({grpc_pick_unused_port_or_die()}); // Enter fallback mode because the LB channel fails to connect. WaitForBackend(0); // Return a new balancer that sends a dead backend. ShutdownBackend(1); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends({{backends_[1]->port()}}), 0); SetNextResolutionForLbChannelAllBalancers(); // The state (TRANSIENT_FAILURE) update from the child policy will be ignored // because we are still in fallback mode. gpr_timespec deadline = gpr_time_add( gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_millis(5000, GPR_TIMESPAN)); // Send 5 seconds worth of RPCs. do { CheckRpcSendOk(); } while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0); // After the backend is restarted, the child policy will eventually be READY, // and we will exit fallback mode. StartBackend(1); WaitForBackend(1); // We have exited fallback mode, so calls will go to the child policy // exclusively. CheckRpcSendOk(100); EXPECT_EQ(0U, backends_[0]->backend_service()->request_count()); EXPECT_EQ(100U, backends_[1]->backend_service()->request_count()); } TEST_F(SingleBalancerTest, BackendsRestart) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(GetBackendPortsInGroups()), 0); WaitForAllBackends(); // Stop backends. RPCs should fail. ShutdownAllBackends(); CheckRpcSendFailure(); // Restart all backends. RPCs should start succeeding again. StartAllBackends(); CheckRpcSendOk(1 /* times */, 2000 /* timeout_ms */, true /* wait_for_ready */); } class UpdatesTest : public XdsEnd2endTest { public: UpdatesTest() : XdsEnd2endTest(4, 3, 0) {} }; TEST_F(UpdatesTest, UpdateBalancersButKeepUsingOriginalBalancer) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); auto first_backend = GetBackendPortsInGroups(0, 1); auto second_backend = GetBackendPortsInGroups(1, 2); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(first_backend), 0); ScheduleResponseForBalancer( 1, EdsServiceImpl::BuildResponseForBackends(second_backend), 0); // Wait until the first backend is ready. WaitForBackend(0); // Send 10 requests. gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // All 10 requests should have gone to the first backend. EXPECT_EQ(10U, backends_[0]->backend_service()->request_count()); // The EDS service of balancer 0 got a single request, and sent a single // response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); gpr_log(GPR_INFO, "========= ABOUT TO UPDATE 1 =========="); SetNextResolutionForLbChannel({balancers_[1]->port()}); gpr_log(GPR_INFO, "========= UPDATE 1 DONE =========="); EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); gpr_timespec deadline = gpr_time_add( gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_millis(10000, GPR_TIMESPAN)); // Send 10 seconds worth of RPCs do { CheckRpcSendOk(); } while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0); // The current LB call is still working, so xds continued using it to the // first balancer, which doesn't assign the second backend. EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); } TEST_F(UpdatesTest, UpdateBalancerName) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); auto first_backend = GetBackendPortsInGroups(0, 1); auto second_backend = GetBackendPortsInGroups(1, 2); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(first_backend), 0); ScheduleResponseForBalancer( 1, EdsServiceImpl::BuildResponseForBackends(second_backend), 0); // Wait until the first backend is ready. WaitForBackend(0); // Send 10 requests. gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // All 10 requests should have gone to the first backend. EXPECT_EQ(10U, backends_[0]->backend_service()->request_count()); // The EDS service of balancer 0 got a single request, and sent a single // response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); std::vector ports; ports.emplace_back(balancers_[1]->port()); auto new_lb_channel_response_generator = grpc_core::MakeRefCounted(); SetNextResolutionForLbChannel(ports, nullptr, new_lb_channel_response_generator.get()); gpr_log(GPR_INFO, "========= ABOUT TO UPDATE BALANCER NAME =========="); SetNextResolution({}, "{\n" " \"loadBalancingConfig\":[\n" " { \"does_not_exist\":{} },\n" " { \"xds_experimental\":{ \"balancerName\": " "\"fake:///updated_lb\" } }\n" " ]\n" "}", new_lb_channel_response_generator.get()); gpr_log(GPR_INFO, "========= UPDATED BALANCER NAME =========="); // Wait until update has been processed, as signaled by the second backend // receiving a request. EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); WaitForBackend(1); backends_[1]->backend_service()->ResetCounters(); gpr_log(GPR_INFO, "========= BEFORE SECOND BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH SECOND BATCH =========="); // All 10 requests should have gone to the second backend. EXPECT_EQ(10U, backends_[1]->backend_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(1U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); } // Send an update with the same set of LBs as the one in SetUp() in order to // verify that the LB channel inside xds keeps the initial connection (which // by definition is also present in the update). TEST_F(UpdatesTest, UpdateBalancersRepeated) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannelAllBalancers(); auto first_backend = GetBackendPortsInGroups(0, 1); auto second_backend = GetBackendPortsInGroups(1, 2); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(first_backend), 0); ScheduleResponseForBalancer( 1, EdsServiceImpl::BuildResponseForBackends(second_backend), 0); // Wait until the first backend is ready. WaitForBackend(0); // Send 10 requests. gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // All 10 requests should have gone to the first backend. EXPECT_EQ(10U, backends_[0]->backend_service()->request_count()); // The EDS service of balancer 0 got a single request, and sent a single // response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); std::vector ports; ports.emplace_back(balancers_[0]->port()); ports.emplace_back(balancers_[1]->port()); ports.emplace_back(balancers_[2]->port()); gpr_log(GPR_INFO, "========= ABOUT TO UPDATE 1 =========="); SetNextResolutionForLbChannel(ports); gpr_log(GPR_INFO, "========= UPDATE 1 DONE =========="); EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); gpr_timespec deadline = gpr_time_add( gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_millis(10000, GPR_TIMESPAN)); // Send 10 seconds worth of RPCs do { CheckRpcSendOk(); } while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0); // xds continued using the original LB call to the first balancer, which // doesn't assign the second backend. EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); ports.clear(); ports.emplace_back(balancers_[0]->port()); ports.emplace_back(balancers_[1]->port()); gpr_log(GPR_INFO, "========= ABOUT TO UPDATE 2 =========="); SetNextResolutionForLbChannel(ports); gpr_log(GPR_INFO, "========= UPDATE 2 DONE =========="); EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); deadline = gpr_time_add(gpr_now(GPR_CLOCK_REALTIME), gpr_time_from_millis(10000, GPR_TIMESPAN)); // Send 10 seconds worth of RPCs do { CheckRpcSendOk(); } while (gpr_time_cmp(gpr_now(GPR_CLOCK_REALTIME), deadline) < 0); // xds continued using the original LB call to the first balancer, which // doesn't assign the second backend. EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); } TEST_F(UpdatesTest, UpdateBalancersDeadUpdate) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({balancers_[0]->port()}); auto first_backend = GetBackendPortsInGroups(0, 1); auto second_backend = GetBackendPortsInGroups(1, 2); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends(first_backend), 0); ScheduleResponseForBalancer( 1, EdsServiceImpl::BuildResponseForBackends(second_backend), 0); // Start servers and send 10 RPCs per server. gpr_log(GPR_INFO, "========= BEFORE FIRST BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH FIRST BATCH =========="); // All 10 requests should have gone to the first backend. EXPECT_EQ(10U, backends_[0]->backend_service()->request_count()); // Kill balancer 0 gpr_log(GPR_INFO, "********** ABOUT TO KILL BALANCER 0 *************"); balancers_[0]->Shutdown(); gpr_log(GPR_INFO, "********** KILLED BALANCER 0 *************"); // This is serviced by the existing child policy. gpr_log(GPR_INFO, "========= BEFORE SECOND BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH SECOND BATCH =========="); // All 10 requests should again have gone to the first backend. EXPECT_EQ(20U, backends_[0]->backend_service()->request_count()); EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); // The EDS service of balancer 0 got a single request, and sent a single // response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[1]->eds_service()->response_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); gpr_log(GPR_INFO, "========= ABOUT TO UPDATE 1 =========="); SetNextResolutionForLbChannel({balancers_[1]->port()}); gpr_log(GPR_INFO, "========= UPDATE 1 DONE =========="); // Wait until update has been processed, as signaled by the second backend // receiving a request. In the meantime, the client continues to be serviced // (by the first backend) without interruption. EXPECT_EQ(0U, backends_[1]->backend_service()->request_count()); WaitForBackend(1); // This is serviced by the updated RR policy backends_[1]->backend_service()->ResetCounters(); gpr_log(GPR_INFO, "========= BEFORE THIRD BATCH =========="); CheckRpcSendOk(10); gpr_log(GPR_INFO, "========= DONE WITH THIRD BATCH =========="); // All 10 requests should have gone to the second backend. EXPECT_EQ(10U, backends_[1]->backend_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); // The second balancer, published as part of the first update, may end up // getting two requests (that is, 1 <= #req <= 2) if the LB call retry timer // firing races with the arrival of the update containing the second // balancer. EXPECT_GE(balancers_[1]->eds_service()->request_count(), 1U); EXPECT_GE(balancers_[1]->eds_service()->response_count(), 1U); EXPECT_LE(balancers_[1]->eds_service()->request_count(), 2U); EXPECT_LE(balancers_[1]->eds_service()->response_count(), 2U); EXPECT_EQ(0U, balancers_[2]->eds_service()->request_count()); EXPECT_EQ(0U, balancers_[2]->eds_service()->response_count()); } // The re-resolution tests are deferred because they rely on the fallback mode, // which hasn't been supported. // TODO(juanlishen): Add TEST_F(UpdatesTest, ReresolveDeadBackend). // TODO(juanlishen): Add TEST_F(UpdatesWithClientLoadReportingTest, // ReresolveDeadBalancer) // The drop tests are deferred because the drop handling hasn't been added yet. // TODO(roth): Add TEST_F(SingleBalancerTest, Drop) // TODO(roth): Add TEST_F(SingleBalancerTest, DropAllFirst) // TODO(roth): Add TEST_F(SingleBalancerTest, DropAll) class SingleBalancerWithClientLoadReportingTest : public XdsEnd2endTest { public: SingleBalancerWithClientLoadReportingTest() : XdsEnd2endTest(4, 1, 3) {} }; TEST_F(SingleBalancerWithClientLoadReportingTest, Vanilla) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({balancers_[0]->port()}); const size_t kNumRpcsPerAddress = 100; // TODO(juanlishen): Partition the backends after multiple localities is // tested. ScheduleResponseForBalancer(0, EdsServiceImpl::BuildResponseForBackends( GetBackendPortsInGroups(0, backends_.size())), 0); // Wait until all backends are ready. int num_ok = 0; int num_failure = 0; int num_drops = 0; std::tie(num_ok, num_failure, num_drops) = WaitForAllBackends(); // Send kNumRpcsPerAddress RPCs per server. CheckRpcSendOk(kNumRpcsPerAddress * num_backends_); // Each backend should have gotten 100 requests. for (size_t i = 0; i < backends_.size(); ++i) { EXPECT_EQ(kNumRpcsPerAddress, backends_[i]->backend_service()->request_count()); } // The EDS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->eds_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->eds_service()->response_count()); // The LRS service got a single request, and sent a single response. EXPECT_EQ(1U, balancers_[0]->lrs_service()->request_count()); EXPECT_EQ(1U, balancers_[0]->lrs_service()->response_count()); // The load report received at the balancer should be correct. ClientStats* client_stats = balancers_[0]->lrs_service()->WaitForLoadReport(); EXPECT_EQ(kNumRpcsPerAddress * num_backends_ + num_ok, client_stats->total_successful_requests()); EXPECT_EQ(0U, client_stats->total_requests_in_progress()); EXPECT_EQ(kNumRpcsPerAddress * num_backends_ + num_ok, client_stats->total_issued_requests()); EXPECT_EQ(0U, client_stats->total_error_requests()); EXPECT_EQ(0U, client_stats->total_dropped_requests()); } TEST_F(SingleBalancerWithClientLoadReportingTest, BalancerRestart) { SetNextResolution({}, kDefaultServiceConfig_.c_str()); SetNextResolutionForLbChannel({balancers_[0]->port()}); const size_t kNumBackendsFirstPass = backends_.size() / 2; const size_t kNumBackendsSecondPass = backends_.size() - kNumBackendsFirstPass; ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends( GetBackendPortsInGroups(0, kNumBackendsFirstPass)), 0); // Wait until all backends returned by the balancer are ready. int num_ok = 0; int num_failure = 0; int num_drops = 0; std::tie(num_ok, num_failure, num_drops) = WaitForAllBackends(/* num_requests_multiple_of */ 1, /* start_index */ 0, /* stop_index */ kNumBackendsFirstPass); ClientStats* client_stats = balancers_[0]->lrs_service()->WaitForLoadReport(); EXPECT_EQ(static_cast(num_ok), client_stats->total_successful_requests()); EXPECT_EQ(0U, client_stats->total_requests_in_progress()); EXPECT_EQ(0U, client_stats->total_error_requests()); EXPECT_EQ(0U, client_stats->total_dropped_requests()); // Shut down the balancer. balancers_[0]->Shutdown(); // Send 1 more request per backend. This will continue using the // last serverlist we received from the balancer before it was shut down. ResetBackendCounters(); CheckRpcSendOk(kNumBackendsFirstPass); int num_started = kNumBackendsFirstPass; // Each backend should have gotten 1 request. for (size_t i = 0; i < kNumBackendsFirstPass; ++i) { EXPECT_EQ(1UL, backends_[i]->backend_service()->request_count()); } // Now restart the balancer, this time pointing to the new backends. balancers_[0]->Start(server_host_); ScheduleResponseForBalancer( 0, EdsServiceImpl::BuildResponseForBackends( GetBackendPortsInGroups(kNumBackendsFirstPass)), 0); // Wait for queries to start going to one of the new backends. // This tells us that we're now using the new serverlist. std::tie(num_ok, num_failure, num_drops) = WaitForAllBackends(/* num_requests_multiple_of */ 1, /* start_index */ kNumBackendsFirstPass); num_started += num_ok + num_failure + num_drops; // Send one RPC per backend. CheckRpcSendOk(kNumBackendsSecondPass); num_started += kNumBackendsSecondPass; // Check client stats. client_stats = balancers_[0]->lrs_service()->WaitForLoadReport(); EXPECT_EQ(num_started, client_stats->total_successful_requests()); EXPECT_EQ(0U, client_stats->total_requests_in_progress()); EXPECT_EQ(0U, client_stats->total_error_requests()); EXPECT_EQ(0U, client_stats->total_dropped_requests()); } // TODO(juanlishen): Add TEST_F(SingleBalancerWithClientLoadReportingTest, Drop) } // namespace } // namespace testing } // namespace grpc int main(int argc, char** argv) { grpc::testing::TestEnvironment env(argc, argv); ::testing::InitGoogleTest(&argc, argv); const auto result = RUN_ALL_TESTS(); return result; }