/* * * Copyright 2016, Google Inc. * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions are * met: * * * Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * * Redistributions in binary form must reproduce the above * copyright notice, this list of conditions and the following disclaimer * in the documentation and/or other materials provided with the * distribution. * * Neither the name of Google Inc. nor the names of its * contributors may be used to endorse or promote products derived from * this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. * */ #include "test/core/util/passthru_endpoint.h" #include #include #include #include #include #include #include "src/core/lib/iomgr/sockaddr.h" #include "src/core/lib/slice/slice_internal.h" typedef struct { grpc_endpoint base; double bytes_per_second; grpc_endpoint *wrapped; gpr_timespec last_write; gpr_mu mu; grpc_slice_buffer write_buffer; grpc_slice_buffer writing_buffer; grpc_error *error; bool writing; } trickle_endpoint; static void te_read(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep, grpc_slice_buffer *slices, grpc_closure *cb) { trickle_endpoint *te = (trickle_endpoint *)ep; grpc_endpoint_read(exec_ctx, te->wrapped, slices, cb); } static void te_write(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep, grpc_slice_buffer *slices, grpc_closure *cb) { trickle_endpoint *te = (trickle_endpoint *)ep; for (size_t i = 0; i < slices->count; i++) { grpc_slice_ref_internal(slices->slices[i]); } gpr_mu_lock(&te->mu); if (te->write_buffer.length == 0) { te->last_write = gpr_now(GPR_CLOCK_MONOTONIC); } grpc_slice_buffer_addn(&te->write_buffer, slices->slices, slices->count); grpc_closure_sched(exec_ctx, cb, GRPC_ERROR_REF(te->error)); gpr_mu_unlock(&te->mu); } static grpc_workqueue *te_get_workqueue(grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; return grpc_endpoint_get_workqueue(te->wrapped); } static void te_add_to_pollset(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep, grpc_pollset *pollset) { trickle_endpoint *te = (trickle_endpoint *)ep; grpc_endpoint_add_to_pollset(exec_ctx, te->wrapped, pollset); } static void te_add_to_pollset_set(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep, grpc_pollset_set *pollset_set) { trickle_endpoint *te = (trickle_endpoint *)ep; grpc_endpoint_add_to_pollset_set(exec_ctx, te->wrapped, pollset_set); } static void te_shutdown(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep, grpc_error *why) { trickle_endpoint *te = (trickle_endpoint *)ep; gpr_mu_lock(&te->mu); if (te->error == GRPC_ERROR_NONE) { te->error = GRPC_ERROR_REF(why); } gpr_mu_unlock(&te->mu); grpc_endpoint_shutdown(exec_ctx, te->wrapped, why); } static void te_destroy(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; grpc_endpoint_destroy(exec_ctx, te->wrapped); gpr_mu_destroy(&te->mu); grpc_slice_buffer_destroy_internal(exec_ctx, &te->write_buffer); grpc_slice_buffer_destroy_internal(exec_ctx, &te->writing_buffer); GRPC_ERROR_UNREF(te->error); gpr_free(te); } static grpc_resource_user *te_get_resource_user(grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; return grpc_endpoint_get_resource_user(te->wrapped); } static char *te_get_peer(grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; return grpc_endpoint_get_peer(te->wrapped); } static int te_get_fd(grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; return grpc_endpoint_get_fd(te->wrapped); } static void te_finish_write(grpc_exec_ctx *exec_ctx, void *arg, grpc_error *error) { trickle_endpoint *te = arg; gpr_mu_lock(&te->mu); te->writing = false; grpc_slice_buffer_reset_and_unref(&te->writing_buffer); gpr_mu_unlock(&te->mu); } static const grpc_endpoint_vtable vtable = {te_read, te_write, te_get_workqueue, te_add_to_pollset, te_add_to_pollset_set, te_shutdown, te_destroy, te_get_resource_user, te_get_peer, te_get_fd}; grpc_endpoint *grpc_trickle_endpoint_create(grpc_endpoint *wrap, double bytes_per_second) { trickle_endpoint *te = gpr_malloc(sizeof(*te)); te->base.vtable = &vtable; te->wrapped = wrap; te->bytes_per_second = bytes_per_second; gpr_mu_init(&te->mu); grpc_slice_buffer_init(&te->write_buffer); grpc_slice_buffer_init(&te->writing_buffer); te->error = GRPC_ERROR_NONE; te->writing = false; return &te->base; } static double ts2dbl(gpr_timespec s) { return s.tv_sec + 1e-9 * s.tv_nsec; } size_t grpc_trickle_endpoint_trickle(grpc_exec_ctx *exec_ctx, grpc_endpoint *ep) { trickle_endpoint *te = (trickle_endpoint *)ep; gpr_mu_lock(&te->mu); if (!te->writing && te->write_buffer.length > 0) { gpr_timespec now = gpr_now(GPR_CLOCK_MONOTONIC); double elapsed = ts2dbl(gpr_time_sub(now, te->last_write)); size_t bytes = (size_t)(te->bytes_per_second * elapsed); // gpr_log(GPR_DEBUG, "%lf elapsed --> %" PRIdPTR " bytes", elapsed, bytes); if (bytes > 0) { grpc_slice_buffer_move_first(&te->write_buffer, GPR_MIN(bytes, te->write_buffer.length), &te->writing_buffer); te->writing = true; te->last_write = now; grpc_endpoint_write( exec_ctx, te->wrapped, &te->writing_buffer, grpc_closure_create(te_finish_write, te, grpc_schedule_on_exec_ctx)); } } size_t backlog = te->write_buffer.length; gpr_mu_unlock(&te->mu); return backlog; }