/* * * Copyright 2015 gRPC authors. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. * */ /* This benchmark exists to ensure that the benchmark integration is * working */ #include #include #include #include #include #include "src/core/lib/surface/completion_queue.h" #include "test/core/util/test_config.h" #include "test/cpp/microbenchmarks/helpers.h" #include "test/cpp/util/test_config.h" namespace grpc { namespace testing { static void BM_CreateDestroyCpp(benchmark::State& state) { TrackCounters track_counters; for (auto _ : state) { CompletionQueue cq; } track_counters.Finish(state); } BENCHMARK(BM_CreateDestroyCpp); /* Create cq using a different constructor */ static void BM_CreateDestroyCpp2(benchmark::State& state) { TrackCounters track_counters; for (auto _ : state) { grpc_completion_queue* core_cq = grpc_completion_queue_create_for_next(nullptr); CompletionQueue cq(core_cq); } track_counters.Finish(state); } BENCHMARK(BM_CreateDestroyCpp2); static void BM_CreateDestroyCore(benchmark::State& state) { TrackCounters track_counters; for (auto _ : state) { // TODO(sreek): Templatize this benchmark and pass completion type and // polling type as parameters grpc_completion_queue_destroy( grpc_completion_queue_create_for_next(nullptr)); } track_counters.Finish(state); } BENCHMARK(BM_CreateDestroyCore); static void DoneWithCompletionOnStack(void* /*arg*/, grpc_cq_completion* /*completion*/) {} static void DoneWithCompletionOnHeap(void* /*arg*/, grpc_cq_completion* completion) { delete completion; } class PhonyTag final : public internal::CompletionQueueTag { public: bool FinalizeResult(void** /*tag*/, bool* /*status*/) override { return true; } }; static void BM_Pass1Cpp(benchmark::State& state) { TrackCounters track_counters; CompletionQueue cq; grpc_completion_queue* c_cq = cq.cq(); for (auto _ : state) { grpc_cq_completion completion; PhonyTag phony_tag; grpc_core::ExecCtx exec_ctx; GPR_ASSERT(grpc_cq_begin_op(c_cq, &phony_tag)); grpc_cq_end_op(c_cq, &phony_tag, GRPC_ERROR_NONE, DoneWithCompletionOnStack, nullptr, &completion); void* tag; bool ok; cq.Next(&tag, &ok); } track_counters.Finish(state); } BENCHMARK(BM_Pass1Cpp); static void BM_Pass1Core(benchmark::State& state) { TrackCounters track_counters; // TODO(sreek): Templatize this benchmark and pass polling_type as a param grpc_completion_queue* cq = grpc_completion_queue_create_for_next(nullptr); gpr_timespec deadline = gpr_inf_future(GPR_CLOCK_MONOTONIC); for (auto _ : state) { grpc_cq_completion completion; grpc_core::ExecCtx exec_ctx; GPR_ASSERT(grpc_cq_begin_op(cq, nullptr)); grpc_cq_end_op(cq, nullptr, GRPC_ERROR_NONE, DoneWithCompletionOnStack, nullptr, &completion); grpc_completion_queue_next(cq, deadline, nullptr); } grpc_completion_queue_destroy(cq); track_counters.Finish(state); } BENCHMARK(BM_Pass1Core); static void BM_Pluck1Core(benchmark::State& state) { TrackCounters track_counters; // TODO(sreek): Templatize this benchmark and pass polling_type as a param grpc_completion_queue* cq = grpc_completion_queue_create_for_pluck(nullptr); gpr_timespec deadline = gpr_inf_future(GPR_CLOCK_MONOTONIC); for (auto _ : state) { grpc_cq_completion completion; grpc_core::ExecCtx exec_ctx; GPR_ASSERT(grpc_cq_begin_op(cq, nullptr)); grpc_cq_end_op(cq, nullptr, GRPC_ERROR_NONE, DoneWithCompletionOnStack, nullptr, &completion); grpc_completion_queue_pluck(cq, nullptr, deadline, nullptr); } grpc_completion_queue_destroy(cq); track_counters.Finish(state); } BENCHMARK(BM_Pluck1Core); static void BM_EmptyCore(benchmark::State& state) { TrackCounters track_counters; // TODO(sreek): Templatize this benchmark and pass polling_type as a param grpc_completion_queue* cq = grpc_completion_queue_create_for_next(nullptr); gpr_timespec deadline = gpr_inf_past(GPR_CLOCK_MONOTONIC); for (auto _ : state) { grpc_completion_queue_next(cq, deadline, nullptr); } grpc_completion_queue_destroy(cq); track_counters.Finish(state); } BENCHMARK(BM_EmptyCore); // Helper for tests to shutdown correctly and tersely static void shutdown_and_destroy(grpc_completion_queue* cc) { grpc_completion_queue_shutdown(cc); grpc_completion_queue_destroy(cc); } static gpr_mu shutdown_mu, mu; static gpr_cv shutdown_cv, cv; // Tag completion queue iterate times class TagCallback : public grpc_completion_queue_functor { public: explicit TagCallback(int* iter) : iter_(iter) { functor_run = &TagCallback::Run; inlineable = false; } ~TagCallback() {} static void Run(grpc_completion_queue_functor* cb, int ok) { gpr_mu_lock(&mu); GPR_ASSERT(static_cast(ok)); *static_cast(cb)->iter_ += 1; gpr_cv_signal(&cv); gpr_mu_unlock(&mu); }; private: int* iter_; }; // Check if completion queue is shut down class ShutdownCallback : public grpc_completion_queue_functor { public: explicit ShutdownCallback(bool* done) : done_(done) { functor_run = &ShutdownCallback::Run; inlineable = false; } ~ShutdownCallback() {} static void Run(grpc_completion_queue_functor* cb, int ok) { gpr_mu_lock(&shutdown_mu); *static_cast(cb)->done_ = static_cast(ok); gpr_cv_signal(&shutdown_cv); gpr_mu_unlock(&shutdown_mu); } private: bool* done_; }; static void BM_Callback_CQ_Pass1Core(benchmark::State& state) { TrackCounters track_counters; int iteration = 0, current_iterations = 0; TagCallback tag_cb(&iteration); gpr_mu_init(&mu); gpr_cv_init(&cv); gpr_mu_init(&shutdown_mu); gpr_cv_init(&shutdown_cv); bool got_shutdown = false; ShutdownCallback shutdown_cb(&got_shutdown); // This test with stack-allocated completions only works for non-polling or // EM-polling callback core CQs because otherwise the callback could execute // on another thread after the stack objects here go out of scope. An // alternative would be to synchronize between the benchmark loop and the // callback, but then it would be measuring the overhead of synchronization // rather than the overhead of the completion queue. // For generality, test here with non-polling. grpc_completion_queue_attributes attr; attr.version = 2; attr.cq_completion_type = GRPC_CQ_CALLBACK; attr.cq_polling_type = GRPC_CQ_NON_POLLING; attr.cq_shutdown_cb = &shutdown_cb; grpc_completion_queue* cc = grpc_completion_queue_create( grpc_completion_queue_factory_lookup(&attr), &attr, nullptr); for (auto _ : state) { grpc_core::ApplicationCallbackExecCtx callback_exec_ctx; grpc_core::ExecCtx exec_ctx; grpc_cq_completion completion; GPR_ASSERT(grpc_cq_begin_op(cc, &tag_cb)); grpc_cq_end_op(cc, &tag_cb, GRPC_ERROR_NONE, DoneWithCompletionOnStack, nullptr, &completion); } shutdown_and_destroy(cc); gpr_mu_lock(&mu); current_iterations = static_cast(state.iterations()); while (current_iterations != iteration) { // Wait for all the callbacks to complete. gpr_cv_wait(&cv, &mu, gpr_inf_future(GPR_CLOCK_REALTIME)); } gpr_mu_unlock(&mu); gpr_mu_lock(&shutdown_mu); while (!got_shutdown) { // Wait for the shutdown callback to complete. gpr_cv_wait(&shutdown_cv, &shutdown_mu, gpr_inf_future(GPR_CLOCK_REALTIME)); } gpr_mu_unlock(&shutdown_mu); GPR_ASSERT(got_shutdown); GPR_ASSERT(iteration == static_cast(state.iterations())); track_counters.Finish(state); gpr_cv_destroy(&cv); gpr_mu_destroy(&mu); gpr_cv_destroy(&shutdown_cv); gpr_mu_destroy(&shutdown_mu); } static void BM_Callback_CQ_Pass1CoreHeapCompletion(benchmark::State& state) { TrackCounters track_counters; int iteration = 0, current_iterations = 0; TagCallback tag_cb(&iteration); gpr_mu_init(&mu); gpr_cv_init(&cv); gpr_mu_init(&shutdown_mu); gpr_cv_init(&shutdown_cv); bool got_shutdown = false; ShutdownCallback shutdown_cb(&got_shutdown); grpc_completion_queue* cc = grpc_completion_queue_create_for_callback(&shutdown_cb, nullptr); for (auto _ : state) { grpc_core::ApplicationCallbackExecCtx callback_exec_ctx; grpc_core::ExecCtx exec_ctx; grpc_cq_completion* completion = new grpc_cq_completion; GPR_ASSERT(grpc_cq_begin_op(cc, &tag_cb)); grpc_cq_end_op(cc, &tag_cb, GRPC_ERROR_NONE, DoneWithCompletionOnHeap, nullptr, completion); } shutdown_and_destroy(cc); gpr_mu_lock(&mu); current_iterations = static_cast(state.iterations()); while (current_iterations != iteration) { // Wait for all the callbacks to complete. gpr_cv_wait(&cv, &mu, gpr_inf_future(GPR_CLOCK_REALTIME)); } gpr_mu_unlock(&mu); gpr_mu_lock(&shutdown_mu); while (!got_shutdown) { // Wait for the shutdown callback to complete. gpr_cv_wait(&shutdown_cv, &shutdown_mu, gpr_inf_future(GPR_CLOCK_REALTIME)); } gpr_mu_unlock(&shutdown_mu); GPR_ASSERT(got_shutdown); GPR_ASSERT(iteration == static_cast(state.iterations())); track_counters.Finish(state); gpr_cv_destroy(&cv); gpr_mu_destroy(&mu); gpr_cv_destroy(&shutdown_cv); gpr_mu_destroy(&shutdown_mu); } BENCHMARK(BM_Callback_CQ_Pass1Core); BENCHMARK(BM_Callback_CQ_Pass1CoreHeapCompletion); } // namespace testing } // namespace grpc // Some distros have RunSpecifiedBenchmarks under the benchmark namespace, // and others do not. This allows us to support both modes. namespace benchmark { void RunTheBenchmarksNamespaced() { RunSpecifiedBenchmarks(); } } // namespace benchmark int main(int argc, char** argv) { grpc::testing::TestEnvironment env(argc, argv); LibraryInitializer libInit; ::benchmark::Initialize(&argc, argv); ::grpc::testing::InitTest(&argc, &argv, false); benchmark::RunTheBenchmarksNamespaced(); return 0; }