/* * * Copyright 2015, Google Inc. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following disclaimer * in the documentation and/or other materials provided with the * distribution. * * Neither the name of Google Inc. nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * */ #include #include #include #include #include #include #include #include #include #include #include #include #include "src/core/lib/profiling/timers.h" #include "src/core/lib/support/env.h" #include "src/proto/grpc/testing/services.grpc.pb.h" #include "test/core/util/port.h" #include "test/core/util/test_config.h" #include "test/cpp/qps/driver.h" #include "test/cpp/qps/histogram.h" #include "test/cpp/qps/qps_worker.h" #include "test/cpp/qps/stats.h" using std::list; using std::thread; using std::unique_ptr; using std::deque; using std::vector; namespace grpc { namespace testing { static std::string get_host(const std::string& worker) { char* host; char* port; gpr_split_host_port(worker.c_str(), &host, &port); const string s(host); gpr_free(host); gpr_free(port); return s; } static std::unordered_map> get_hosts_and_cores( const deque& workers) { std::unordered_map> hosts; for (auto it = workers.begin(); it != workers.end(); it++) { const string host = get_host(*it); if (hosts.find(host) == hosts.end()) { auto stub = WorkerService::NewStub( CreateChannel(*it, InsecureChannelCredentials())); grpc::ClientContext ctx; ctx.set_wait_for_ready(true); CoreRequest dummy; CoreResponse cores; grpc::Status s = stub->CoreCount(&ctx, dummy, &cores); GPR_ASSERT(s.ok()); std::deque dq; for (int i = 0; i < cores.cores(); i++) { dq.push_back(i); } hosts[host] = dq; } } return hosts; } static deque get_workers(const string& name) { char* env = gpr_getenv(name.c_str()); if (!env) return deque(); deque out; char* p = env; for (;;) { char* comma = strchr(p, ','); if (comma) { out.emplace_back(p, comma); p = comma + 1; } else { out.emplace_back(p); gpr_free(env); return out; } } } // helpers for postprocess_scenario_result static double WallTime(ClientStats s) { return s.time_elapsed(); } static double SystemTime(ClientStats s) { return s.time_system(); } static double UserTime(ClientStats s) { return s.time_user(); } static double ServerWallTime(ServerStats s) { return s.time_elapsed(); } static double ServerSystemTime(ServerStats s) { return s.time_system(); } static double ServerUserTime(ServerStats s) { return s.time_user(); } static int Cores(int n) { return n; } // Postprocess ScenarioResult and populate result summary. static void postprocess_scenario_result(ScenarioResult* result) { Histogram histogram; histogram.MergeProto(result->latencies()); auto time_estimate = average(result->client_stats(), WallTime); auto qps = histogram.Count() / time_estimate; auto qps_per_server_core = qps / sum(result->server_cores(), Cores); result->mutable_summary()->set_qps(qps); result->mutable_summary()->set_qps_per_server_core(qps_per_server_core); result->mutable_summary()->set_latency_50(histogram.Percentile(50)); result->mutable_summary()->set_latency_90(histogram.Percentile(90)); result->mutable_summary()->set_latency_95(histogram.Percentile(95)); result->mutable_summary()->set_latency_99(histogram.Percentile(99)); result->mutable_summary()->set_latency_999(histogram.Percentile(99.9)); auto server_system_time = 100.0 * sum(result->server_stats(), ServerSystemTime) / sum(result->server_stats(), ServerWallTime); auto server_user_time = 100.0 * sum(result->server_stats(), ServerUserTime) / sum(result->server_stats(), ServerWallTime); auto client_system_time = 100.0 * sum(result->client_stats(), SystemTime) / sum(result->client_stats(), WallTime); auto client_user_time = 100.0 * sum(result->client_stats(), UserTime) / sum(result->client_stats(), WallTime); result->mutable_summary()->set_server_system_time(server_system_time); result->mutable_summary()->set_server_user_time(server_user_time); result->mutable_summary()->set_client_system_time(client_system_time); result->mutable_summary()->set_client_user_time(client_user_time); if (result->request_results_size() > 0) { int64_t successes = 0; int64_t failures = 0; for (int i = 0; i < result->request_results_size(); i++) { RequestResultCount rrc = result->request_results(i); if (rrc.status_code() == 0) { successes += rrc.count(); } else { failures += rrc.count(); } } result->mutable_summary()->set_successful_requests_per_second(successes / time_estimate); result->mutable_summary()->set_failed_requests_per_second(failures / time_estimate); } } // Namespace for classes and functions used only in RunScenario // Using this rather than local definitions to workaround gcc-4.4 limitations // regarding using templates without linkage namespace runsc { // ClientContext allocator static ClientContext* AllocContext(list* contexts) { contexts->emplace_back(); auto context = &contexts->back(); context->set_wait_for_ready(true); return context; } struct ServerData { unique_ptr stub; unique_ptr> stream; }; struct ClientData { unique_ptr stub; unique_ptr> stream; }; } // namespace runsc std::unique_ptr RunScenario( const ClientConfig& initial_client_config, size_t num_clients, const ServerConfig& initial_server_config, size_t num_servers, int warmup_seconds, int benchmark_seconds, int spawn_local_worker_count) { // Log everything from the driver gpr_set_log_verbosity(GPR_LOG_SEVERITY_DEBUG); // ClientContext allocations (all are destroyed at scope exit) list contexts; // To be added to the result, containing the final configuration used for // client and config (including host, etc.) ClientConfig result_client_config; const ServerConfig result_server_config = initial_server_config; // Get client, server lists auto workers = get_workers("QPS_WORKERS"); ClientConfig client_config = initial_client_config; // Spawn some local workers if desired vector> local_workers; for (int i = 0; i < abs(spawn_local_worker_count); i++) { // act as if we're a new test -- gets a good rng seed static bool called_init = false; if (!called_init) { char args_buf[100]; strcpy(args_buf, "some-benchmark"); char* args[] = {args_buf}; grpc_test_init(1, args); called_init = true; } int driver_port = grpc_pick_unused_port_or_die(); local_workers.emplace_back(new QpsWorker(driver_port)); char addr[256]; sprintf(addr, "localhost:%d", driver_port); if (spawn_local_worker_count < 0) { workers.push_front(addr); } else { workers.push_back(addr); } } // Setup the hosts and core counts auto hosts_cores = get_hosts_and_cores(workers); // if num_clients is set to <=0, do dynamic sizing: all workers // except for servers are clients if (num_clients <= 0) { num_clients = workers.size() - num_servers; } // TODO(ctiller): support running multiple configurations, and binpack // client/server pairs // to available workers GPR_ASSERT(workers.size() >= num_clients + num_servers); // Trim to just what we need workers.resize(num_clients + num_servers); // Start servers using runsc::ServerData; // servers is array rather than std::vector to avoid gcc-4.4 issues // where class contained in std::vector must have a copy constructor auto* servers = new ServerData[num_servers]; for (size_t i = 0; i < num_servers; i++) { gpr_log(GPR_INFO, "Starting server on %s (worker #%" PRIuPTR ")", workers[i].c_str(), i); servers[i].stub = WorkerService::NewStub( CreateChannel(workers[i], InsecureChannelCredentials())); ServerConfig server_config = initial_server_config; char* host; char* driver_port; char* cli_target; gpr_split_host_port(workers[i].c_str(), &host, &driver_port); string host_str(host); int server_core_limit = initial_server_config.core_limit(); int client_core_limit = initial_client_config.core_limit(); if (server_core_limit == 0 && client_core_limit > 0) { // In this case, limit the server cores if it matches the // same host as one or more clients const auto& dq = hosts_cores.at(host_str); bool match = false; int limit = dq.size(); for (size_t cli = 0; cli < num_clients; cli++) { if (host_str == get_host(workers[cli + num_servers])) { limit -= client_core_limit; match = true; } } if (match) { GPR_ASSERT(limit > 0); server_core_limit = limit; } } if (server_core_limit > 0) { auto& dq = hosts_cores.at(host_str); GPR_ASSERT(dq.size() >= static_cast(server_core_limit)); for (int core = 0; core < server_core_limit; core++) { server_config.add_core_list(dq.front()); dq.pop_front(); } } ServerArgs args; *args.mutable_setup() = server_config; servers[i].stream = servers[i].stub->RunServer(runsc::AllocContext(&contexts)); if (!servers[i].stream->Write(args)) { gpr_log(GPR_ERROR, "Could not write args to server %zu", i); } ServerStatus init_status; if (!servers[i].stream->Read(&init_status)) { gpr_log(GPR_ERROR, "Server %zu did not yield initial status", i); } gpr_join_host_port(&cli_target, host, init_status.port()); client_config.add_server_targets(cli_target); gpr_free(host); gpr_free(driver_port); gpr_free(cli_target); } // Targets are all set by now result_client_config = client_config; // Start clients using runsc::ClientData; // clients is array rather than std::vector to avoid gcc-4.4 issues // where class contained in std::vector must have a copy constructor auto* clients = new ClientData[num_clients]; size_t channels_allocated = 0; for (size_t i = 0; i < num_clients; i++) { const auto& worker = workers[i + num_servers]; gpr_log(GPR_INFO, "Starting client on %s (worker #%" PRIuPTR ")", worker.c_str(), i + num_servers); clients[i].stub = WorkerService::NewStub( CreateChannel(worker, InsecureChannelCredentials())); ClientConfig per_client_config = client_config; int server_core_limit = initial_server_config.core_limit(); int client_core_limit = initial_client_config.core_limit(); if ((server_core_limit > 0) || (client_core_limit > 0)) { auto& dq = hosts_cores.at(get_host(worker)); if (client_core_limit == 0) { // limit client cores if it matches a server host bool match = false; int limit = dq.size(); for (size_t srv = 0; srv < num_servers; srv++) { if (get_host(worker) == get_host(workers[srv])) { match = true; } } if (match) { GPR_ASSERT(limit > 0); client_core_limit = limit; } } if (client_core_limit > 0) { GPR_ASSERT(dq.size() >= static_cast(client_core_limit)); for (int core = 0; core < client_core_limit; core++) { per_client_config.add_core_list(dq.front()); dq.pop_front(); } } } // Reduce channel count so that total channels specified is held regardless // of the number of clients available size_t num_channels = (client_config.client_channels() - channels_allocated) / (num_clients - i); channels_allocated += num_channels; gpr_log(GPR_DEBUG, "Client %" PRIdPTR " gets %" PRIdPTR " channels", i, num_channels); per_client_config.set_client_channels(num_channels); ClientArgs args; *args.mutable_setup() = per_client_config; clients[i].stream = clients[i].stub->RunClient(runsc::AllocContext(&contexts)); if (!clients[i].stream->Write(args)) { gpr_log(GPR_ERROR, "Could not write args to client %zu", i); } } for (size_t i = 0; i < num_clients; i++) { ClientStatus init_status; if (!clients[i].stream->Read(&init_status)) { gpr_log(GPR_ERROR, "Client %zu did not yield initial status", i); } } // Send an initial mark: clients can use this to know that everything is ready // to start gpr_log(GPR_INFO, "Initiating"); ServerArgs server_mark; server_mark.mutable_mark()->set_reset(true); ClientArgs client_mark; client_mark.mutable_mark()->set_reset(true); ServerStatus server_status; ClientStatus client_status; for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; if (!client->stream->Write(client_mark)) { gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i); } } for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; if (!client->stream->Read(&client_status)) { gpr_log(GPR_ERROR, "Couldn't get status from client %zu", i); } } // Let everything warmup gpr_log(GPR_INFO, "Warming up"); gpr_timespec start = gpr_now(GPR_CLOCK_REALTIME); gpr_sleep_until( gpr_time_add(start, gpr_time_from_seconds(warmup_seconds, GPR_TIMESPAN))); // Start a run gpr_log(GPR_INFO, "Starting"); for (size_t i = 0; i < num_servers; i++) { auto server = &servers[i]; if (!server->stream->Write(server_mark)) { gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i); } } for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; if (!client->stream->Write(client_mark)) { gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i); } } for (size_t i = 0; i < num_servers; i++) { auto server = &servers[i]; if (!server->stream->Read(&server_status)) { gpr_log(GPR_ERROR, "Couldn't get status from server %zu", i); } } for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; if (!client->stream->Read(&client_status)) { gpr_log(GPR_ERROR, "Couldn't get status from client %zu", i); } } // Wait some time gpr_log(GPR_INFO, "Running"); // Use gpr_sleep_until rather than this_thread::sleep_until to support // compilers that don't work with this_thread gpr_sleep_until(gpr_time_add( start, gpr_time_from_seconds(warmup_seconds + benchmark_seconds, GPR_TIMESPAN))); gpr_timer_set_enabled(0); // Finish a run std::unique_ptr result(new ScenarioResult); Histogram merged_latencies; std::unordered_map merged_statuses; gpr_log(GPR_INFO, "Finishing clients"); for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; if (!client->stream->Write(client_mark)) { gpr_log(GPR_ERROR, "Couldn't write mark to client %zu", i); } if (!client->stream->WritesDone()) { gpr_log(GPR_ERROR, "Failed WritesDone for client %zu", i); } } for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; // Read the client final status if (client->stream->Read(&client_status)) { gpr_log(GPR_INFO, "Received final status from client %zu", i); const auto& stats = client_status.stats(); merged_latencies.MergeProto(stats.latencies()); for (int i = 0; i < stats.request_results_size(); i++) { merged_statuses[stats.request_results(i).status_code()] += stats.request_results(i).count(); } result->add_client_stats()->CopyFrom(stats); // That final status should be the last message on the client stream GPR_ASSERT(!client->stream->Read(&client_status)); } else { gpr_log(GPR_ERROR, "Couldn't get final status from client %zu", i); } } for (size_t i = 0; i < num_clients; i++) { auto client = &clients[i]; Status s = client->stream->Finish(); result->add_client_success(s.ok()); if (!s.ok()) { gpr_log(GPR_ERROR, "Client %zu had an error %s", i, s.error_message().c_str()); } } delete[] clients; merged_latencies.FillProto(result->mutable_latencies()); for (std::unordered_map::iterator it = merged_statuses.begin(); it != merged_statuses.end(); ++it) { RequestResultCount* rrc = result->add_request_results(); rrc->set_status_code(it->first); rrc->set_count(it->second); } gpr_log(GPR_INFO, "Finishing servers"); for (size_t i = 0; i < num_servers; i++) { auto server = &servers[i]; if (!server->stream->Write(server_mark)) { gpr_log(GPR_ERROR, "Couldn't write mark to server %zu", i); } if (!server->stream->WritesDone()) { gpr_log(GPR_ERROR, "Failed WritesDone for server %zu", i); } } for (size_t i = 0; i < num_servers; i++) { auto server = &servers[i]; // Read the server final status if (server->stream->Read(&server_status)) { gpr_log(GPR_INFO, "Received final status from server %zu", i); result->add_server_stats()->CopyFrom(server_status.stats()); result->add_server_cores(server_status.cores()); // That final status should be the last message on the server stream GPR_ASSERT(!server->stream->Read(&server_status)); } else { gpr_log(GPR_ERROR, "Couldn't get final status from server %zu", i); } } for (size_t i = 0; i < num_servers; i++) { auto server = &servers[i]; Status s = server->stream->Finish(); result->add_server_success(s.ok()); if (!s.ok()) { gpr_log(GPR_ERROR, "Server %zu had an error %s", i, s.error_message().c_str()); } } delete[] servers; postprocess_scenario_result(result.get()); return result; } bool RunQuit() { // Get client, server lists bool result = true; auto workers = get_workers("QPS_WORKERS"); for (size_t i = 0; i < workers.size(); i++) { auto stub = WorkerService::NewStub( CreateChannel(workers[i], InsecureChannelCredentials())); Void dummy; grpc::ClientContext ctx; ctx.set_wait_for_ready(true); Status s = stub->QuitWorker(&ctx, dummy, &dummy); if (!s.ok()) { gpr_log(GPR_ERROR, "Worker %zu could not be properly quit because %s", i, s.error_message().c_str()); result = false; } } return result; } } // namespace testing } // namespace grpc