|
|
|
@ -796,7 +796,8 @@ static void cc_destroy_channel_elem(grpc_exec_ctx *exec_ctx, |
|
|
|
|
// send_message
|
|
|
|
|
// recv_trailing_metadata
|
|
|
|
|
// send_trailing_metadata
|
|
|
|
|
#define MAX_WAITING_BATCHES 6 |
|
|
|
|
// We also add room for a single cancel_stream batch.
|
|
|
|
|
#define MAX_WAITING_BATCHES 7 |
|
|
|
|
|
|
|
|
|
/** Call data. Holds a pointer to grpc_subchannel_call and the
|
|
|
|
|
associated machinery to create such a pointer. |
|
|
|
@ -808,23 +809,25 @@ typedef struct client_channel_call_data { |
|
|
|
|
// The code in deadline_filter.c requires this to be the first field.
|
|
|
|
|
// TODO(roth): This is slightly sub-optimal in that grpc_deadline_state
|
|
|
|
|
// and this struct both independently store a pointer to the call
|
|
|
|
|
// stack and each has its own mutex. If/when we have time, find a way
|
|
|
|
|
// to avoid this without breaking the grpc_deadline_state abstraction.
|
|
|
|
|
// combiner. If/when we have time, find a way to avoid this without
|
|
|
|
|
// breaking the grpc_deadline_state abstraction.
|
|
|
|
|
grpc_deadline_state deadline_state; |
|
|
|
|
|
|
|
|
|
grpc_slice path; // Request path.
|
|
|
|
|
gpr_timespec call_start_time; |
|
|
|
|
gpr_timespec deadline; |
|
|
|
|
gpr_arena *arena; |
|
|
|
|
grpc_call_combiner *call_combiner; |
|
|
|
|
|
|
|
|
|
grpc_server_retry_throttle_data *retry_throttle_data; |
|
|
|
|
method_parameters *method_params; |
|
|
|
|
|
|
|
|
|
/** either 0 for no call, a pointer to a grpc_subchannel_call (if the lowest
|
|
|
|
|
bit is 0), or a pointer to an error (if the lowest bit is 1) */ |
|
|
|
|
gpr_atm subchannel_call_or_error; |
|
|
|
|
gpr_arena *arena; |
|
|
|
|
grpc_subchannel_call *subchannel_call; |
|
|
|
|
grpc_error *error; |
|
|
|
|
|
|
|
|
|
grpc_lb_policy *lb_policy; // Holds ref while LB pick is pending.
|
|
|
|
|
grpc_closure lb_pick_closure; |
|
|
|
|
grpc_closure cancel_closure; |
|
|
|
|
|
|
|
|
|
grpc_connected_subchannel *connected_subchannel; |
|
|
|
|
grpc_call_context_element subchannel_call_context[GRPC_CONTEXT_COUNT]; |
|
|
|
@ -832,10 +835,9 @@ typedef struct client_channel_call_data { |
|
|
|
|
|
|
|
|
|
grpc_transport_stream_op_batch *waiting_for_pick_batches[MAX_WAITING_BATCHES]; |
|
|
|
|
size_t waiting_for_pick_batches_count; |
|
|
|
|
grpc_closure handle_pending_batch_in_call_combiner[MAX_WAITING_BATCHES]; |
|
|
|
|
|
|
|
|
|
grpc_transport_stream_op_batch_payload *initial_metadata_payload; |
|
|
|
|
|
|
|
|
|
grpc_call_stack *owning_call; |
|
|
|
|
grpc_transport_stream_op_batch *initial_metadata_batch; |
|
|
|
|
|
|
|
|
|
grpc_linked_mdelem lb_token_mdelem; |
|
|
|
|
|
|
|
|
@ -843,53 +845,40 @@ typedef struct client_channel_call_data { |
|
|
|
|
grpc_closure *original_on_complete; |
|
|
|
|
} call_data; |
|
|
|
|
|
|
|
|
|
typedef struct { |
|
|
|
|
grpc_subchannel_call *subchannel_call; |
|
|
|
|
grpc_error *error; |
|
|
|
|
} call_or_error; |
|
|
|
|
|
|
|
|
|
static call_or_error get_call_or_error(call_data *p) { |
|
|
|
|
gpr_atm c = gpr_atm_acq_load(&p->subchannel_call_or_error); |
|
|
|
|
if (c == 0) |
|
|
|
|
return (call_or_error){NULL, NULL}; |
|
|
|
|
else if (c & 1) |
|
|
|
|
return (call_or_error){NULL, (grpc_error *)((c) & ~(gpr_atm)1)}; |
|
|
|
|
else |
|
|
|
|
return (call_or_error){(grpc_subchannel_call *)c, NULL}; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static bool set_call_or_error(call_data *p, call_or_error coe) { |
|
|
|
|
// this should always be under a lock
|
|
|
|
|
call_or_error existing = get_call_or_error(p); |
|
|
|
|
if (existing.error != GRPC_ERROR_NONE) { |
|
|
|
|
GRPC_ERROR_UNREF(coe.error); |
|
|
|
|
return false; |
|
|
|
|
} |
|
|
|
|
GPR_ASSERT(existing.subchannel_call == NULL); |
|
|
|
|
if (coe.error != GRPC_ERROR_NONE) { |
|
|
|
|
GPR_ASSERT(coe.subchannel_call == NULL); |
|
|
|
|
gpr_atm_rel_store(&p->subchannel_call_or_error, 1 | (gpr_atm)coe.error); |
|
|
|
|
} else { |
|
|
|
|
GPR_ASSERT(coe.subchannel_call != NULL); |
|
|
|
|
gpr_atm_rel_store(&p->subchannel_call_or_error, |
|
|
|
|
(gpr_atm)coe.subchannel_call); |
|
|
|
|
} |
|
|
|
|
return true; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
grpc_subchannel_call *grpc_client_channel_get_subchannel_call( |
|
|
|
|
grpc_call_element *call_elem) { |
|
|
|
|
return get_call_or_error(call_elem->call_data).subchannel_call; |
|
|
|
|
grpc_call_element *elem) { |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
return calld->subchannel_call; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void waiting_for_pick_batches_add_locked( |
|
|
|
|
// This is called via the call combiner, so access to calld is synchronized.
|
|
|
|
|
static void waiting_for_pick_batches_add( |
|
|
|
|
call_data *calld, grpc_transport_stream_op_batch *batch) { |
|
|
|
|
if (batch->send_initial_metadata) { |
|
|
|
|
GPR_ASSERT(calld->initial_metadata_batch == NULL); |
|
|
|
|
calld->initial_metadata_batch = batch; |
|
|
|
|
} else { |
|
|
|
|
GPR_ASSERT(calld->waiting_for_pick_batches_count < MAX_WAITING_BATCHES); |
|
|
|
|
calld->waiting_for_pick_batches[calld->waiting_for_pick_batches_count++] = |
|
|
|
|
batch; |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void waiting_for_pick_batches_fail_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
// This is called via the call combiner, so access to calld is synchronized.
|
|
|
|
|
static void fail_pending_batch_in_call_combiner(grpc_exec_ctx *exec_ctx, |
|
|
|
|
void *arg, grpc_error *error) { |
|
|
|
|
call_data *calld = arg; |
|
|
|
|
if (calld->waiting_for_pick_batches_count > 0) { |
|
|
|
|
--calld->waiting_for_pick_batches_count; |
|
|
|
|
grpc_transport_stream_op_batch_finish_with_failure( |
|
|
|
|
exec_ctx, |
|
|
|
|
calld->waiting_for_pick_batches[calld->waiting_for_pick_batches_count], |
|
|
|
|
GRPC_ERROR_REF(error), calld->call_combiner); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// This is called via the call combiner, so access to calld is synchronized.
|
|
|
|
|
static void waiting_for_pick_batches_fail(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
@ -900,34 +889,60 @@ static void waiting_for_pick_batches_fail_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_error_string(error)); |
|
|
|
|
} |
|
|
|
|
for (size_t i = 0; i < calld->waiting_for_pick_batches_count; ++i) { |
|
|
|
|
GRPC_CLOSURE_INIT(&calld->handle_pending_batch_in_call_combiner[i], |
|
|
|
|
fail_pending_batch_in_call_combiner, calld, |
|
|
|
|
grpc_schedule_on_exec_ctx); |
|
|
|
|
GRPC_CALL_COMBINER_START(exec_ctx, calld->call_combiner, |
|
|
|
|
&calld->handle_pending_batch_in_call_combiner[i], |
|
|
|
|
GRPC_ERROR_REF(error), |
|
|
|
|
"waiting_for_pick_batches_fail"); |
|
|
|
|
} |
|
|
|
|
if (calld->initial_metadata_batch != NULL) { |
|
|
|
|
grpc_transport_stream_op_batch_finish_with_failure( |
|
|
|
|
exec_ctx, calld->waiting_for_pick_batches[i], GRPC_ERROR_REF(error)); |
|
|
|
|
exec_ctx, calld->initial_metadata_batch, GRPC_ERROR_REF(error), |
|
|
|
|
calld->call_combiner); |
|
|
|
|
} else { |
|
|
|
|
GRPC_CALL_COMBINER_STOP(exec_ctx, calld->call_combiner, |
|
|
|
|
"waiting_for_pick_batches_fail"); |
|
|
|
|
} |
|
|
|
|
calld->waiting_for_pick_batches_count = 0; |
|
|
|
|
GRPC_ERROR_UNREF(error); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void waiting_for_pick_batches_resume_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
// This is called via the call combiner, so access to calld is synchronized.
|
|
|
|
|
static void run_pending_batch_in_call_combiner(grpc_exec_ctx *exec_ctx, |
|
|
|
|
void *arg, grpc_error *ignored) { |
|
|
|
|
call_data *calld = arg; |
|
|
|
|
if (calld->waiting_for_pick_batches_count > 0) { |
|
|
|
|
--calld->waiting_for_pick_batches_count; |
|
|
|
|
grpc_subchannel_call_process_op( |
|
|
|
|
exec_ctx, calld->subchannel_call, |
|
|
|
|
calld->waiting_for_pick_batches[calld->waiting_for_pick_batches_count]); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// This is called via the call combiner, so access to calld is synchronized.
|
|
|
|
|
static void waiting_for_pick_batches_resume(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem) { |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
if (calld->waiting_for_pick_batches_count == 0) return; |
|
|
|
|
call_or_error coe = get_call_or_error(calld); |
|
|
|
|
if (coe.error != GRPC_ERROR_NONE) { |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_REF(coe.error)); |
|
|
|
|
return; |
|
|
|
|
} |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: sending %" PRIdPTR |
|
|
|
|
" pending batches to subchannel_call=%p", |
|
|
|
|
elem->channel_data, calld, calld->waiting_for_pick_batches_count, |
|
|
|
|
coe.subchannel_call); |
|
|
|
|
chand, calld, calld->waiting_for_pick_batches_count, |
|
|
|
|
calld->subchannel_call); |
|
|
|
|
} |
|
|
|
|
for (size_t i = 0; i < calld->waiting_for_pick_batches_count; ++i) { |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, coe.subchannel_call, |
|
|
|
|
calld->waiting_for_pick_batches[i]); |
|
|
|
|
} |
|
|
|
|
calld->waiting_for_pick_batches_count = 0; |
|
|
|
|
GRPC_CLOSURE_INIT(&calld->handle_pending_batch_in_call_combiner[i], |
|
|
|
|
run_pending_batch_in_call_combiner, calld, |
|
|
|
|
grpc_schedule_on_exec_ctx); |
|
|
|
|
GRPC_CALL_COMBINER_START(exec_ctx, calld->call_combiner, |
|
|
|
|
&calld->handle_pending_batch_in_call_combiner[i], |
|
|
|
|
GRPC_ERROR_NONE, |
|
|
|
|
"waiting_for_pick_batches_resume"); |
|
|
|
|
} |
|
|
|
|
GPR_ASSERT(calld->initial_metadata_batch != NULL); |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, calld->subchannel_call, |
|
|
|
|
calld->initial_metadata_batch); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// Applies service config to the call. Must be invoked once we know
|
|
|
|
@ -968,29 +983,28 @@ static void apply_service_config_to_call_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
static void create_subchannel_call_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
grpc_subchannel_call *subchannel_call = NULL; |
|
|
|
|
const grpc_connected_subchannel_call_args call_args = { |
|
|
|
|
.pollent = calld->pollent, |
|
|
|
|
.path = calld->path, |
|
|
|
|
.start_time = calld->call_start_time, |
|
|
|
|
.deadline = calld->deadline, |
|
|
|
|
.arena = calld->arena, |
|
|
|
|
.context = calld->subchannel_call_context}; |
|
|
|
|
.context = calld->subchannel_call_context, |
|
|
|
|
.call_combiner = calld->call_combiner}; |
|
|
|
|
grpc_error *new_error = grpc_connected_subchannel_create_call( |
|
|
|
|
exec_ctx, calld->connected_subchannel, &call_args, &subchannel_call); |
|
|
|
|
exec_ctx, calld->connected_subchannel, &call_args, |
|
|
|
|
&calld->subchannel_call); |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: create subchannel_call=%p: error=%s", |
|
|
|
|
elem->channel_data, calld, subchannel_call, |
|
|
|
|
grpc_error_string(new_error)); |
|
|
|
|
chand, calld, calld->subchannel_call, grpc_error_string(new_error)); |
|
|
|
|
} |
|
|
|
|
GPR_ASSERT(set_call_or_error( |
|
|
|
|
calld, (call_or_error){.subchannel_call = subchannel_call})); |
|
|
|
|
if (new_error != GRPC_ERROR_NONE) { |
|
|
|
|
new_error = grpc_error_add_child(new_error, error); |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, new_error); |
|
|
|
|
waiting_for_pick_batches_fail(exec_ctx, elem, new_error); |
|
|
|
|
} else { |
|
|
|
|
waiting_for_pick_batches_resume_locked(exec_ctx, elem); |
|
|
|
|
waiting_for_pick_batches_resume(exec_ctx, elem); |
|
|
|
|
} |
|
|
|
|
GRPC_ERROR_UNREF(error); |
|
|
|
|
} |
|
|
|
@ -1002,11 +1016,10 @@ static void subchannel_ready_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
grpc_polling_entity_del_from_pollset_set(exec_ctx, calld->pollent, |
|
|
|
|
chand->interested_parties); |
|
|
|
|
call_or_error coe = get_call_or_error(calld); |
|
|
|
|
if (calld->connected_subchannel == NULL) { |
|
|
|
|
// Failed to create subchannel.
|
|
|
|
|
grpc_error *failure = |
|
|
|
|
error == GRPC_ERROR_NONE |
|
|
|
|
GRPC_ERROR_UNREF(calld->error); |
|
|
|
|
calld->error = error == GRPC_ERROR_NONE |
|
|
|
|
? GRPC_ERROR_CREATE_FROM_STATIC_STRING( |
|
|
|
|
"Call dropped by load balancing policy") |
|
|
|
|
: GRPC_ERROR_CREATE_REFERENCING_FROM_STATIC_STRING( |
|
|
|
@ -1014,48 +1027,16 @@ static void subchannel_ready_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: failed to create subchannel: error=%s", chand, |
|
|
|
|
calld, grpc_error_string(failure)); |
|
|
|
|
} |
|
|
|
|
set_call_or_error(calld, (call_or_error){.error = GRPC_ERROR_REF(failure)}); |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, failure); |
|
|
|
|
} else if (coe.error != GRPC_ERROR_NONE) { |
|
|
|
|
/* already cancelled before subchannel became ready */ |
|
|
|
|
grpc_error *child_errors[] = {error, coe.error}; |
|
|
|
|
grpc_error *cancellation_error = |
|
|
|
|
GRPC_ERROR_CREATE_REFERENCING_FROM_STATIC_STRING( |
|
|
|
|
"Cancelled before creating subchannel", child_errors, |
|
|
|
|
GPR_ARRAY_SIZE(child_errors)); |
|
|
|
|
/* if due to deadline, attach the deadline exceeded status to the error */ |
|
|
|
|
if (gpr_time_cmp(calld->deadline, gpr_now(GPR_CLOCK_MONOTONIC)) < 0) { |
|
|
|
|
cancellation_error = |
|
|
|
|
grpc_error_set_int(cancellation_error, GRPC_ERROR_INT_GRPC_STATUS, |
|
|
|
|
GRPC_STATUS_DEADLINE_EXCEEDED); |
|
|
|
|
calld, grpc_error_string(calld->error)); |
|
|
|
|
} |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: cancelled before subchannel became ready: %s", |
|
|
|
|
chand, calld, grpc_error_string(cancellation_error)); |
|
|
|
|
} |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, cancellation_error); |
|
|
|
|
waiting_for_pick_batches_fail(exec_ctx, elem, GRPC_ERROR_REF(calld->error)); |
|
|
|
|
} else { |
|
|
|
|
/* Create call on subchannel. */ |
|
|
|
|
create_subchannel_call_locked(exec_ctx, elem, GRPC_ERROR_REF(error)); |
|
|
|
|
} |
|
|
|
|
GRPC_CALL_STACK_UNREF(exec_ctx, calld->owning_call, "pick_subchannel"); |
|
|
|
|
GRPC_ERROR_UNREF(error); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static char *cc_get_peer(grpc_exec_ctx *exec_ctx, grpc_call_element *elem) { |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
grpc_subchannel_call *subchannel_call = |
|
|
|
|
get_call_or_error(calld).subchannel_call; |
|
|
|
|
if (subchannel_call == NULL) { |
|
|
|
|
return NULL; |
|
|
|
|
} else { |
|
|
|
|
return grpc_subchannel_call_get_peer(exec_ctx, subchannel_call); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/** Return true if subchannel is available immediately (in which case
|
|
|
|
|
subchannel_ready_locked() should not be called), or false otherwise (in |
|
|
|
|
which case subchannel_ready_locked() should be called when the subchannel |
|
|
|
@ -1069,6 +1050,44 @@ typedef struct { |
|
|
|
|
grpc_closure closure; |
|
|
|
|
} pick_after_resolver_result_args; |
|
|
|
|
|
|
|
|
|
// Note: This runs under the client_channel combiner, but will NOT be
|
|
|
|
|
// holding the call combiner.
|
|
|
|
|
static void pick_after_resolver_result_cancel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
void *arg, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
grpc_call_element *elem = arg; |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
// If we don't yet have a resolver result, then a closure for
|
|
|
|
|
// pick_after_resolver_result_done_locked() will have been added to
|
|
|
|
|
// chand->waiting_for_resolver_result_closures, and it may not be invoked
|
|
|
|
|
// until after this call has been destroyed. We mark the operation as
|
|
|
|
|
// cancelled, so that when pick_after_resolver_result_done_locked()
|
|
|
|
|
// is called, it will be a no-op. We also immediately invoke
|
|
|
|
|
// subchannel_ready_locked() to propagate the error back to the caller.
|
|
|
|
|
for (grpc_closure *closure = chand->waiting_for_resolver_result_closures.head; |
|
|
|
|
closure != NULL; closure = closure->next_data.next) { |
|
|
|
|
pick_after_resolver_result_args *args = closure->cb_arg; |
|
|
|
|
if (!args->cancelled && args->elem == elem) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: " |
|
|
|
|
"cancelling pick waiting for resolver result", |
|
|
|
|
chand, calld); |
|
|
|
|
} |
|
|
|
|
args->cancelled = true; |
|
|
|
|
// Note: Although we are not in the call combiner here, we are
|
|
|
|
|
// basically stealing the call combiner from the pending pick, so
|
|
|
|
|
// it's safe to call subchannel_ready_locked() here -- we are
|
|
|
|
|
// essentially calling it here instead of calling it in
|
|
|
|
|
// pick_after_resolver_result_done_locked().
|
|
|
|
|
subchannel_ready_locked(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_CREATE_REFERENCING_FROM_STATIC_STRING( |
|
|
|
|
"Pick cancelled", &error, 1)); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void pick_after_resolver_result_done_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
void *arg, |
|
|
|
|
grpc_error *error) { |
|
|
|
@ -1079,21 +1098,24 @@ static void pick_after_resolver_result_done_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
gpr_log(GPR_DEBUG, "call cancelled before resolver result"); |
|
|
|
|
} |
|
|
|
|
} else { |
|
|
|
|
channel_data *chand = args->elem->channel_data; |
|
|
|
|
call_data *calld = args->elem->call_data; |
|
|
|
|
grpc_call_element *elem = args->elem; |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
grpc_call_combiner_set_notify_on_cancel(exec_ctx, calld->call_combiner, |
|
|
|
|
NULL); |
|
|
|
|
if (error != GRPC_ERROR_NONE) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: resolver failed to return data", |
|
|
|
|
chand, calld); |
|
|
|
|
} |
|
|
|
|
subchannel_ready_locked(exec_ctx, args->elem, GRPC_ERROR_REF(error)); |
|
|
|
|
subchannel_ready_locked(exec_ctx, elem, GRPC_ERROR_REF(error)); |
|
|
|
|
} else { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: resolver returned, doing pick", |
|
|
|
|
chand, calld); |
|
|
|
|
} |
|
|
|
|
if (pick_subchannel_locked(exec_ctx, args->elem)) { |
|
|
|
|
subchannel_ready_locked(exec_ctx, args->elem, GRPC_ERROR_NONE); |
|
|
|
|
if (pick_subchannel_locked(exec_ctx, elem)) { |
|
|
|
|
subchannel_ready_locked(exec_ctx, elem, GRPC_ERROR_NONE); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
@ -1116,41 +1138,33 @@ static void pick_after_resolver_result_start_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
args, grpc_combiner_scheduler(chand->combiner)); |
|
|
|
|
grpc_closure_list_append(&chand->waiting_for_resolver_result_closures, |
|
|
|
|
&args->closure, GRPC_ERROR_NONE); |
|
|
|
|
grpc_call_combiner_set_notify_on_cancel( |
|
|
|
|
exec_ctx, calld->call_combiner, |
|
|
|
|
GRPC_CLOSURE_INIT(&calld->cancel_closure, |
|
|
|
|
pick_after_resolver_result_cancel_locked, elem, |
|
|
|
|
grpc_combiner_scheduler(chand->combiner))); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void pick_after_resolver_result_cancel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem, |
|
|
|
|
// Note: This runs under the client_channel combiner, but will NOT be
|
|
|
|
|
// holding the call combiner.
|
|
|
|
|
static void pick_callback_cancel_locked(grpc_exec_ctx *exec_ctx, void *arg, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
grpc_call_element *elem = arg; |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
// If we don't yet have a resolver result, then a closure for
|
|
|
|
|
// pick_after_resolver_result_done_locked() will have been added to
|
|
|
|
|
// chand->waiting_for_resolver_result_closures, and it may not be invoked
|
|
|
|
|
// until after this call has been destroyed. We mark the operation as
|
|
|
|
|
// cancelled, so that when pick_after_resolver_result_done_locked()
|
|
|
|
|
// is called, it will be a no-op. We also immediately invoke
|
|
|
|
|
// subchannel_ready_locked() to propagate the error back to the caller.
|
|
|
|
|
for (grpc_closure *closure = chand->waiting_for_resolver_result_closures.head; |
|
|
|
|
closure != NULL; closure = closure->next_data.next) { |
|
|
|
|
pick_after_resolver_result_args *args = closure->cb_arg; |
|
|
|
|
if (!args->cancelled && args->elem == elem) { |
|
|
|
|
if (calld->lb_policy != NULL) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: " |
|
|
|
|
"cancelling pick waiting for resolver result", |
|
|
|
|
chand, calld); |
|
|
|
|
} |
|
|
|
|
args->cancelled = true; |
|
|
|
|
subchannel_ready_locked(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_CREATE_REFERENCING_FROM_STATIC_STRING( |
|
|
|
|
"Pick cancelled", &error, 1)); |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: cancelling pick from LB policy %p", |
|
|
|
|
chand, calld, calld->lb_policy); |
|
|
|
|
} |
|
|
|
|
grpc_lb_policy_cancel_pick_locked(exec_ctx, calld->lb_policy, |
|
|
|
|
&calld->connected_subchannel, |
|
|
|
|
GRPC_ERROR_REF(error)); |
|
|
|
|
} |
|
|
|
|
GRPC_ERROR_UNREF(error); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
// Callback invoked by grpc_lb_policy_pick_locked() for async picks.
|
|
|
|
|
// Unrefs the LB policy after invoking subchannel_ready_locked().
|
|
|
|
|
// Unrefs the LB policy and invokes subchannel_ready_locked().
|
|
|
|
|
static void pick_callback_done_locked(grpc_exec_ctx *exec_ctx, void *arg, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
grpc_call_element *elem = arg; |
|
|
|
@ -1160,6 +1174,7 @@ static void pick_callback_done_locked(grpc_exec_ctx *exec_ctx, void *arg, |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: pick completed asynchronously", |
|
|
|
|
chand, calld); |
|
|
|
|
} |
|
|
|
|
grpc_call_combiner_set_notify_on_cancel(exec_ctx, calld->call_combiner, NULL); |
|
|
|
|
GPR_ASSERT(calld->lb_policy != NULL); |
|
|
|
|
GRPC_LB_POLICY_UNREF(exec_ctx, calld->lb_policy, "pick_subchannel"); |
|
|
|
|
calld->lb_policy = NULL; |
|
|
|
@ -1194,24 +1209,15 @@ static bool pick_callback_start_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
} |
|
|
|
|
GRPC_LB_POLICY_UNREF(exec_ctx, calld->lb_policy, "pick_subchannel"); |
|
|
|
|
calld->lb_policy = NULL; |
|
|
|
|
} else { |
|
|
|
|
grpc_call_combiner_set_notify_on_cancel( |
|
|
|
|
exec_ctx, calld->call_combiner, |
|
|
|
|
GRPC_CLOSURE_INIT(&calld->cancel_closure, pick_callback_cancel_locked, |
|
|
|
|
elem, grpc_combiner_scheduler(chand->combiner))); |
|
|
|
|
} |
|
|
|
|
return pick_done; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void pick_callback_cancel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem, |
|
|
|
|
grpc_error *error) { |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
GPR_ASSERT(calld->lb_policy != NULL); |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: cancelling pick from LB policy %p", |
|
|
|
|
chand, calld, calld->lb_policy); |
|
|
|
|
} |
|
|
|
|
grpc_lb_policy_cancel_pick_locked(exec_ctx, calld->lb_policy, |
|
|
|
|
&calld->connected_subchannel, error); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static bool pick_subchannel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
grpc_call_element *elem) { |
|
|
|
|
GPR_TIMER_BEGIN("pick_subchannel", 0); |
|
|
|
@ -1224,7 +1230,7 @@ static bool pick_subchannel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
// Otherwise, if the service config specified a value for this
|
|
|
|
|
// method, use that.
|
|
|
|
|
uint32_t initial_metadata_flags = |
|
|
|
|
calld->initial_metadata_payload->send_initial_metadata |
|
|
|
|
calld->initial_metadata_batch->payload->send_initial_metadata |
|
|
|
|
.send_initial_metadata_flags; |
|
|
|
|
const bool wait_for_ready_set_from_api = |
|
|
|
|
initial_metadata_flags & |
|
|
|
@ -1241,7 +1247,7 @@ static bool pick_subchannel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
const grpc_lb_policy_pick_args inputs = { |
|
|
|
|
calld->initial_metadata_payload->send_initial_metadata |
|
|
|
|
calld->initial_metadata_batch->payload->send_initial_metadata |
|
|
|
|
.send_initial_metadata, |
|
|
|
|
initial_metadata_flags, &calld->lb_token_mdelem}; |
|
|
|
|
pick_done = pick_callback_start_locked(exec_ctx, elem, &inputs); |
|
|
|
@ -1258,91 +1264,33 @@ static bool pick_subchannel_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
return pick_done; |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void start_transport_stream_op_batch_locked(grpc_exec_ctx *exec_ctx, |
|
|
|
|
void *arg, |
|
|
|
|
static void start_pick_locked(grpc_exec_ctx *exec_ctx, void *arg, |
|
|
|
|
grpc_error *error_ignored) { |
|
|
|
|
GPR_TIMER_BEGIN("start_transport_stream_op_batch_locked", 0); |
|
|
|
|
grpc_transport_stream_op_batch *batch = arg; |
|
|
|
|
grpc_call_element *elem = batch->handler_private.extra_arg; |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
/* need to recheck that another thread hasn't set the call */ |
|
|
|
|
call_or_error coe = get_call_or_error(calld); |
|
|
|
|
if (coe.error != GRPC_ERROR_NONE) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: failing batch with error: %s", |
|
|
|
|
chand, calld, grpc_error_string(coe.error)); |
|
|
|
|
} |
|
|
|
|
grpc_transport_stream_op_batch_finish_with_failure( |
|
|
|
|
exec_ctx, batch, GRPC_ERROR_REF(coe.error)); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
if (coe.subchannel_call != NULL) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: sending batch to subchannel_call=%p", chand, |
|
|
|
|
calld, coe.subchannel_call); |
|
|
|
|
} |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, coe.subchannel_call, batch); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
// Add to waiting-for-pick list. If we succeed in getting a
|
|
|
|
|
// subchannel call below, we'll handle this batch (along with any
|
|
|
|
|
// other waiting batches) in waiting_for_pick_batches_resume_locked().
|
|
|
|
|
waiting_for_pick_batches_add_locked(calld, batch); |
|
|
|
|
// If this is a cancellation, cancel the pending pick (if any) and
|
|
|
|
|
// fail any pending batches.
|
|
|
|
|
if (batch->cancel_stream) { |
|
|
|
|
grpc_error *error = batch->payload->cancel_stream.cancel_error; |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: recording cancel_error=%s", chand, |
|
|
|
|
calld, grpc_error_string(error)); |
|
|
|
|
} |
|
|
|
|
/* Stash a copy of cancel_error in our call data, so that we can use
|
|
|
|
|
it for subsequent operations. This ensures that if the call is |
|
|
|
|
cancelled before any batches are passed down (e.g., if the deadline |
|
|
|
|
is in the past when the call starts), we can return the right |
|
|
|
|
error to the caller when the first batch does get passed down. */ |
|
|
|
|
set_call_or_error(calld, (call_or_error){.error = GRPC_ERROR_REF(error)}); |
|
|
|
|
if (calld->lb_policy != NULL) { |
|
|
|
|
pick_callback_cancel_locked(exec_ctx, elem, GRPC_ERROR_REF(error)); |
|
|
|
|
} else { |
|
|
|
|
pick_after_resolver_result_cancel_locked(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_REF(error)); |
|
|
|
|
} |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, GRPC_ERROR_REF(error)); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
/* if we don't have a subchannel, try to get one */ |
|
|
|
|
if (batch->send_initial_metadata) { |
|
|
|
|
GPR_TIMER_BEGIN("start_pick_locked", 0); |
|
|
|
|
grpc_call_element *elem = (grpc_call_element *)arg; |
|
|
|
|
call_data *calld = (call_data *)elem->call_data; |
|
|
|
|
channel_data *chand = (channel_data *)elem->channel_data; |
|
|
|
|
GPR_ASSERT(calld->connected_subchannel == NULL); |
|
|
|
|
calld->initial_metadata_payload = batch->payload; |
|
|
|
|
GRPC_CALL_STACK_REF(calld->owning_call, "pick_subchannel"); |
|
|
|
|
/* If a subchannel is not available immediately, the polling entity from
|
|
|
|
|
call_data should be provided to channel_data's interested_parties, so |
|
|
|
|
that IO of the lb_policy and resolver could be done under it. */ |
|
|
|
|
if (pick_subchannel_locked(exec_ctx, elem)) { |
|
|
|
|
// Pick was returned synchronously.
|
|
|
|
|
GRPC_CALL_STACK_UNREF(exec_ctx, calld->owning_call, "pick_subchannel"); |
|
|
|
|
if (calld->connected_subchannel == NULL) { |
|
|
|
|
grpc_error *error = GRPC_ERROR_CREATE_FROM_STATIC_STRING( |
|
|
|
|
GRPC_ERROR_UNREF(calld->error); |
|
|
|
|
calld->error = GRPC_ERROR_CREATE_FROM_STATIC_STRING( |
|
|
|
|
"Call dropped by load balancing policy"); |
|
|
|
|
set_call_or_error(calld, |
|
|
|
|
(call_or_error){.error = GRPC_ERROR_REF(error)}); |
|
|
|
|
waiting_for_pick_batches_fail_locked(exec_ctx, elem, error); |
|
|
|
|
waiting_for_pick_batches_fail(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_REF(calld->error)); |
|
|
|
|
} else { |
|
|
|
|
// Create subchannel call.
|
|
|
|
|
create_subchannel_call_locked(exec_ctx, elem, GRPC_ERROR_NONE); |
|
|
|
|
} |
|
|
|
|
} else { |
|
|
|
|
// Pick will be done asynchronously. Add the call's polling entity to
|
|
|
|
|
// the channel's interested_parties, so that I/O for the resolver
|
|
|
|
|
// and LB policy can be done under it.
|
|
|
|
|
grpc_polling_entity_add_to_pollset_set(exec_ctx, calld->pollent, |
|
|
|
|
chand->interested_parties); |
|
|
|
|
} |
|
|
|
|
} |
|
|
|
|
done: |
|
|
|
|
GRPC_CALL_STACK_UNREF(exec_ctx, calld->owning_call, |
|
|
|
|
"start_transport_stream_op_batch"); |
|
|
|
|
GPR_TIMER_END("start_transport_stream_op_batch_locked", 0); |
|
|
|
|
GPR_TIMER_END("start_pick_locked", 0); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
static void on_complete(grpc_exec_ctx *exec_ctx, void *arg, grpc_error *error) { |
|
|
|
@ -1365,27 +1313,49 @@ static void on_complete(grpc_exec_ctx *exec_ctx, void *arg, grpc_error *error) { |
|
|
|
|
GRPC_ERROR_REF(error)); |
|
|
|
|
} |
|
|
|
|
|
|
|
|
|
/* The logic here is fairly complicated, due to (a) the fact that we
|
|
|
|
|
need to handle the case where we receive the send op before the |
|
|
|
|
initial metadata op, and (b) the need for efficiency, especially in |
|
|
|
|
the streaming case. |
|
|
|
|
|
|
|
|
|
We use double-checked locking to initially see if initialization has been |
|
|
|
|
performed. If it has not, we acquire the combiner and perform initialization. |
|
|
|
|
If it has, we proceed on the fast path. */ |
|
|
|
|
static void cc_start_transport_stream_op_batch( |
|
|
|
|
grpc_exec_ctx *exec_ctx, grpc_call_element *elem, |
|
|
|
|
grpc_transport_stream_op_batch *batch) { |
|
|
|
|
call_data *calld = elem->call_data; |
|
|
|
|
channel_data *chand = elem->channel_data; |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace) || |
|
|
|
|
GRPC_TRACER_ON(grpc_trace_channel)) { |
|
|
|
|
grpc_call_log_op(GPR_INFO, elem, batch); |
|
|
|
|
} |
|
|
|
|
if (chand->deadline_checking_enabled) { |
|
|
|
|
grpc_deadline_state_client_start_transport_stream_op_batch(exec_ctx, elem, |
|
|
|
|
batch); |
|
|
|
|
} |
|
|
|
|
GPR_TIMER_BEGIN("cc_start_transport_stream_op_batch", 0); |
|
|
|
|
// If we've previously been cancelled, immediately fail any new batches.
|
|
|
|
|
if (calld->error != GRPC_ERROR_NONE) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: failing batch with error: %s", |
|
|
|
|
chand, calld, grpc_error_string(calld->error)); |
|
|
|
|
} |
|
|
|
|
grpc_transport_stream_op_batch_finish_with_failure( |
|
|
|
|
exec_ctx, batch, GRPC_ERROR_REF(calld->error), calld->call_combiner); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
if (batch->cancel_stream) { |
|
|
|
|
// Stash a copy of cancel_error in our call data, so that we can use
|
|
|
|
|
// it for subsequent operations. This ensures that if the call is
|
|
|
|
|
// cancelled before any batches are passed down (e.g., if the deadline
|
|
|
|
|
// is in the past when the call starts), we can return the right
|
|
|
|
|
// error to the caller when the first batch does get passed down.
|
|
|
|
|
GRPC_ERROR_UNREF(calld->error); |
|
|
|
|
calld->error = GRPC_ERROR_REF(batch->payload->cancel_stream.cancel_error); |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: recording cancel_error=%s", chand, |
|
|
|
|
calld, grpc_error_string(calld->error)); |
|
|
|
|
} |
|
|
|
|
// If we have a subchannel call, send the cancellation batch down.
|
|
|
|
|
// Otherwise, fail all pending batches.
|
|
|
|
|
if (calld->subchannel_call != NULL) { |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, calld->subchannel_call, batch); |
|
|
|
|
} else { |
|
|
|
|
waiting_for_pick_batches_add(calld, batch); |
|
|
|
|
waiting_for_pick_batches_fail(exec_ctx, elem, |
|
|
|
|
GRPC_ERROR_REF(calld->error)); |
|
|
|
|
} |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
// Intercept on_complete for recv_trailing_metadata so that we can
|
|
|
|
|
// check retry throttle status.
|
|
|
|
|
if (batch->recv_trailing_metadata) { |
|
|
|
@ -1395,38 +1365,43 @@ static void cc_start_transport_stream_op_batch( |
|
|
|
|
grpc_schedule_on_exec_ctx); |
|
|
|
|
batch->on_complete = &calld->on_complete; |
|
|
|
|
} |
|
|
|
|
/* try to (atomically) get the call */ |
|
|
|
|
call_or_error coe = get_call_or_error(calld); |
|
|
|
|
GPR_TIMER_BEGIN("cc_start_transport_stream_op_batch", 0); |
|
|
|
|
if (coe.error != GRPC_ERROR_NONE) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: failing batch with error: %s", |
|
|
|
|
chand, calld, grpc_error_string(coe.error)); |
|
|
|
|
} |
|
|
|
|
grpc_transport_stream_op_batch_finish_with_failure( |
|
|
|
|
exec_ctx, batch, GRPC_ERROR_REF(coe.error)); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
if (coe.subchannel_call != NULL) { |
|
|
|
|
// Check if we've already gotten a subchannel call.
|
|
|
|
|
// Note that once we have completed the pick, we do not need to enter
|
|
|
|
|
// the channel combiner, which is more efficient (especially for
|
|
|
|
|
// streaming calls).
|
|
|
|
|
if (calld->subchannel_call != NULL) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: sending batch to subchannel_call=%p", chand, |
|
|
|
|
calld, coe.subchannel_call); |
|
|
|
|
calld, calld->subchannel_call); |
|
|
|
|
} |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, coe.subchannel_call, batch); |
|
|
|
|
grpc_subchannel_call_process_op(exec_ctx, calld->subchannel_call, batch); |
|
|
|
|
goto done; |
|
|
|
|
} |
|
|
|
|
/* we failed; lock and figure out what to do */ |
|
|
|
|
// We do not yet have a subchannel call.
|
|
|
|
|
// Add the batch to the waiting-for-pick list.
|
|
|
|
|
waiting_for_pick_batches_add(calld, batch); |
|
|
|
|
// For batches containing a send_initial_metadata op, enter the channel
|
|
|
|
|
// combiner to start a pick.
|
|
|
|
|
if (batch->send_initial_metadata) { |
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, "chand=%p calld=%p: entering combiner", chand, calld); |
|
|
|
|
} |
|
|
|
|
GRPC_CALL_STACK_REF(calld->owning_call, "start_transport_stream_op_batch"); |
|
|
|
|
batch->handler_private.extra_arg = elem; |
|
|
|
|
GRPC_CLOSURE_SCHED( |
|
|
|
|
exec_ctx, GRPC_CLOSURE_INIT(&batch->handler_private.closure, |
|
|
|
|
start_transport_stream_op_batch_locked, batch, |
|
|
|
|
grpc_combiner_scheduler(chand->combiner)), |
|
|
|
|
exec_ctx, |
|
|
|
|
GRPC_CLOSURE_INIT(&batch->handler_private.closure, start_pick_locked, |
|
|
|
|
elem, grpc_combiner_scheduler(chand->combiner)), |
|
|
|
|
GRPC_ERROR_NONE); |
|
|
|
|
} else { |
|
|
|
|
// For all other batches, release the call combiner.
|
|
|
|
|
if (GRPC_TRACER_ON(grpc_client_channel_trace)) { |
|
|
|
|
gpr_log(GPR_DEBUG, |
|
|
|
|
"chand=%p calld=%p: saved batch, yeilding call combiner", chand, |
|
|
|
|
calld); |
|
|
|
|
} |
|
|
|
|
GRPC_CALL_COMBINER_STOP(exec_ctx, calld->call_combiner, |
|
|
|
|
"batch does not include send_initial_metadata"); |
|
|
|
|
} |
|
|
|
|
done: |
|
|
|
|
GPR_TIMER_END("cc_start_transport_stream_op_batch", 0); |
|
|
|
|
} |
|
|
|
@ -1441,10 +1416,11 @@ static grpc_error *cc_init_call_elem(grpc_exec_ctx *exec_ctx, |
|
|
|
|
calld->path = grpc_slice_ref_internal(args->path); |
|
|
|
|
calld->call_start_time = args->start_time; |
|
|
|
|
calld->deadline = gpr_convert_clock_type(args->deadline, GPR_CLOCK_MONOTONIC); |
|
|
|
|
calld->owning_call = args->call_stack; |
|
|
|
|
calld->arena = args->arena; |
|
|
|
|
calld->call_combiner = args->call_combiner; |
|
|
|
|
if (chand->deadline_checking_enabled) { |
|
|
|
|
grpc_deadline_state_init(exec_ctx, elem, args->call_stack, calld->deadline); |
|
|
|
|
grpc_deadline_state_init(exec_ctx, elem, args->call_stack, |
|
|
|
|
args->call_combiner, calld->deadline); |
|
|
|
|
} |
|
|
|
|
return GRPC_ERROR_NONE; |
|
|
|
|
} |
|
|
|
@ -1463,13 +1439,12 @@ static void cc_destroy_call_elem(grpc_exec_ctx *exec_ctx, |
|
|
|
|
if (calld->method_params != NULL) { |
|
|
|
|
method_parameters_unref(calld->method_params); |
|
|
|
|
} |
|
|
|
|
call_or_error coe = get_call_or_error(calld); |
|
|
|
|
GRPC_ERROR_UNREF(coe.error); |
|
|
|
|
if (coe.subchannel_call != NULL) { |
|
|
|
|
grpc_subchannel_call_set_cleanup_closure(coe.subchannel_call, |
|
|
|
|
GRPC_ERROR_UNREF(calld->error); |
|
|
|
|
if (calld->subchannel_call != NULL) { |
|
|
|
|
grpc_subchannel_call_set_cleanup_closure(calld->subchannel_call, |
|
|
|
|
then_schedule_closure); |
|
|
|
|
then_schedule_closure = NULL; |
|
|
|
|
GRPC_SUBCHANNEL_CALL_UNREF(exec_ctx, coe.subchannel_call, |
|
|
|
|
GRPC_SUBCHANNEL_CALL_UNREF(exec_ctx, calld->subchannel_call, |
|
|
|
|
"client_channel_destroy_call"); |
|
|
|
|
} |
|
|
|
|
GPR_ASSERT(calld->lb_policy == NULL); |
|
|
|
@ -1508,7 +1483,6 @@ const grpc_channel_filter grpc_client_channel_filter = { |
|
|
|
|
sizeof(channel_data), |
|
|
|
|
cc_init_channel_elem, |
|
|
|
|
cc_destroy_channel_elem, |
|
|
|
|
cc_get_peer, |
|
|
|
|
cc_get_channel_info, |
|
|
|
|
"client-channel", |
|
|
|
|
}; |
|
|
|
|