mirror of https://github.com/grpc/grpc.git
commit
88966ab145
441 changed files with 12527 additions and 5104 deletions
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,97 @@ |
|||||||
|
Load Balancing in gRPC |
||||||
|
======================= |
||||||
|
|
||||||
|
# Objective |
||||||
|
|
||||||
|
To design a load balancing API between a gRPC client and a Load Balancer to |
||||||
|
instruct the client how to send load to multiple backend servers. |
||||||
|
|
||||||
|
# Background |
||||||
|
|
||||||
|
Prior to any gRPC specifics, we explore some usual ways to approach load |
||||||
|
balancing. |
||||||
|
|
||||||
|
### Proxy Model |
||||||
|
|
||||||
|
Using a proxy provides a solid trustable client that can report load to the load |
||||||
|
balancing system. Proxies typically require more resources to operate since they |
||||||
|
have temporary copies of the RPC request and response. This model also increases |
||||||
|
latency to the RPCs. |
||||||
|
|
||||||
|
The proxy model was deemed inefficient when considering request heavy services |
||||||
|
like storage. |
||||||
|
|
||||||
|
### Balancing-aware Client |
||||||
|
|
||||||
|
This thicker client places more of the load balancing logic in the client. For |
||||||
|
example, the client could contain many load balancing policies (Round Robin, |
||||||
|
Random, etc) used to select servers from a list. In this model, a list of |
||||||
|
servers would be either statically configured in the client, provided by the |
||||||
|
name resolution system, an external load balancer, etc. In any case, the client |
||||||
|
is responsible for choosing the preferred server from the list. |
||||||
|
|
||||||
|
One of the drawbacks of this approach is writing and maintaining the load |
||||||
|
balancing policies in multiple languages and/or versions of the clients. These |
||||||
|
policies can be fairly complicated. Some of the algorithms also require client |
||||||
|
to server communication so the client would need to get thicker to support |
||||||
|
additional RPCs to get health or load information in addition to sending RPCs |
||||||
|
for user requests. |
||||||
|
|
||||||
|
It would also significantly complicate the client's code: the new design hides |
||||||
|
the load balancing complexity of multiple layers and presents it as a simple |
||||||
|
list of servers to the client. |
||||||
|
|
||||||
|
### External Load Balancing Service |
||||||
|
|
||||||
|
The client load balancing code is kept simple and portable, implementing |
||||||
|
well-known algorithms (ie, Round Robin) for server selection. |
||||||
|
Complex load balancing algorithms are instead provided by the load balancer. The |
||||||
|
client relies on the load balancer to provide _load balancing configuration_ and |
||||||
|
_the list of servers_ to which the client should send requests. The balancer |
||||||
|
updates the server list as needed to balance the load as well as handle server |
||||||
|
unavailability or health issues. The load balancer will make any necessary |
||||||
|
complex decisions and inform the client. The load balancer may communicate with |
||||||
|
the backend servers to collect load and health information. |
||||||
|
|
||||||
|
# Proposed Architecture |
||||||
|
|
||||||
|
The gRPC load balancing approach follows the third approach, by having an |
||||||
|
external load balancer which provides simple clients with a list of servers. |
||||||
|
|
||||||
|
## Client |
||||||
|
|
||||||
|
When establishing a gRPC stream to the balancer, the client will send an initial |
||||||
|
request to the load balancer (via a regular gRPC message). The load balancer |
||||||
|
will respond with client config (including, for example, settings for flow |
||||||
|
control, RPC deadlines, etc.) or a redirect to another load balancer. If the |
||||||
|
balancer did not redirect the client, it will then send a list of servers to the |
||||||
|
client. The client will contain simple load balancing logic for choosing the |
||||||
|
next server when it needs to send a request. |
||||||
|
|
||||||
|
## Load Balancer |
||||||
|
|
||||||
|
The Load Balancer is responsible for providing the client with a list of servers |
||||||
|
and client RPC parameters. The balancer chooses when to update the list of |
||||||
|
servers and can decide whether to provide a complete list, a subset, or a |
||||||
|
specific list of “picked” servers in a particular order. The balancer can |
||||||
|
optionally provide an expiration interval after which the server list should no |
||||||
|
longer be trusted and should be updated by the balancer. |
||||||
|
|
||||||
|
The load balancer may open reporting streams to each server contained in the |
||||||
|
server list. These streams are primarily used for load reporting. For example, |
||||||
|
Weighted Round Robin requires that the servers report utilization to the load |
||||||
|
balancer in order to compute the next list of servers. |
||||||
|
|
||||||
|
## Server |
||||||
|
|
||||||
|
The gRPC Server is responsible for answering RPC requests and providing |
||||||
|
responses to the client. The server will also report load to the load balancer |
||||||
|
if a reporting stream was opened for this purpose. |
||||||
|
|
||||||
|
### Security |
||||||
|
|
||||||
|
The load balancer may be separate from the actual server backends and a |
||||||
|
compromise of the load balancer should only lead to a compromise of the |
||||||
|
loadbalancing functionality. In other words, a compromised load balancer should |
||||||
|
not be able to cause a client to trust a (potentially malicious) backend server |
||||||
|
any more than in a comparable situation without loadbalancing. |
@ -0,0 +1,600 @@ |
|||||||
|
/*
|
||||||
|
* |
||||||
|
* Copyright 2015-2016, Google Inc. |
||||||
|
* All rights reserved. |
||||||
|
* |
||||||
|
* Redistribution and use in source and binary forms, with or without |
||||||
|
* modification, are permitted provided that the following conditions are |
||||||
|
* met: |
||||||
|
* |
||||||
|
* * Redistributions of source code must retain the above copyright |
||||||
|
* notice, this list of conditions and the following disclaimer. |
||||||
|
* * Redistributions in binary form must reproduce the above |
||||||
|
* copyright notice, this list of conditions and the following disclaimer |
||||||
|
* in the documentation and/or other materials provided with the |
||||||
|
* distribution. |
||||||
|
* * Neither the name of Google Inc. nor the names of its |
||||||
|
* contributors may be used to endorse or promote products derived from |
||||||
|
* this software without specific prior written permission. |
||||||
|
* |
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
||||||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
||||||
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
||||||
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
||||||
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||||||
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||||||
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||||||
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||||||
|
* |
||||||
|
*/ |
||||||
|
|
||||||
|
// Implements an efficient in-memory log, optimized for multiple writers and
|
||||||
|
// a single reader. Available log space is divided up in blocks of
|
||||||
|
// CENSUS_LOG_2_MAX_RECORD_SIZE bytes. A block can be in one of the following
|
||||||
|
// three data structures:
|
||||||
|
// - Free blocks (free_block_list)
|
||||||
|
// - Blocks with unread data (dirty_block_list)
|
||||||
|
// - Blocks currently attached to cores (core_local_blocks[])
|
||||||
|
//
|
||||||
|
// census_log_start_write() moves a block from core_local_blocks[] to the end of
|
||||||
|
// dirty_block_list when block:
|
||||||
|
// - is out-of-space OR
|
||||||
|
// - has an incomplete record (an incomplete record occurs when a thread calls
|
||||||
|
// census_log_start_write() and is context-switched before calling
|
||||||
|
// census_log_end_write()
|
||||||
|
// So, blocks in dirty_block_list are ordered, from oldest to newest, by the
|
||||||
|
// time when block is detached from the core.
|
||||||
|
//
|
||||||
|
// census_log_read_next() first iterates over dirty_block_list and then
|
||||||
|
// core_local_blocks[]. It moves completely read blocks from dirty_block_list
|
||||||
|
// to free_block_list. Blocks in core_local_blocks[] are not freed, even when
|
||||||
|
// completely read.
|
||||||
|
//
|
||||||
|
// If the log is configured to discard old records and free_block_list is empty,
|
||||||
|
// census_log_start_write() iterates over dirty_block_list to allocate a
|
||||||
|
// new block. It moves the oldest available block (no pending read/write) to
|
||||||
|
// core_local_blocks[].
|
||||||
|
//
|
||||||
|
// core_local_block_struct is used to implement a map from core id to the block
|
||||||
|
// associated with that core. This mapping is advisory. It is possible that the
|
||||||
|
// block returned by this mapping is no longer associated with that core. This
|
||||||
|
// mapping is updated, lazily, by census_log_start_write().
|
||||||
|
//
|
||||||
|
// Locking in block struct:
|
||||||
|
//
|
||||||
|
// Exclusive g_log.lock must be held before calling any functions operating on
|
||||||
|
// block structs except census_log_start_write() and census_log_end_write().
|
||||||
|
//
|
||||||
|
// Writes to a block are serialized via writer_lock. census_log_start_write()
|
||||||
|
// acquires this lock and census_log_end_write() releases it. On failure to
|
||||||
|
// acquire the lock, writer allocates a new block for the current core and
|
||||||
|
// updates core_local_block accordingly.
|
||||||
|
//
|
||||||
|
// Simultaneous read and write access is allowed. Readers can safely read up to
|
||||||
|
// committed bytes (bytes_committed).
|
||||||
|
//
|
||||||
|
// reader_lock protects the block, currently being read, from getting recycled.
|
||||||
|
// start_read() acquires reader_lock and end_read() releases the lock.
|
||||||
|
//
|
||||||
|
// Read/write access to a block is disabled via try_disable_access(). It returns
|
||||||
|
// with both writer_lock and reader_lock held. These locks are subsequently
|
||||||
|
// released by enable_access() to enable access to the block.
|
||||||
|
//
|
||||||
|
// A note on naming: Most function/struct names are prepended by cl_
|
||||||
|
// (shorthand for census_log). Further, functions that manipulate structures
|
||||||
|
// include the name of the structure, which will be passed as the first
|
||||||
|
// argument. E.g. cl_block_initialize() will initialize a cl_block.
|
||||||
|
|
||||||
|
#include "src/core/census/mlog.h" |
||||||
|
#include <grpc/support/alloc.h> |
||||||
|
#include <grpc/support/atm.h> |
||||||
|
#include <grpc/support/cpu.h> |
||||||
|
#include <grpc/support/log.h> |
||||||
|
#include <grpc/support/sync.h> |
||||||
|
#include <grpc/support/useful.h> |
||||||
|
#include <stdbool.h> |
||||||
|
#include <string.h> |
||||||
|
|
||||||
|
// End of platform specific code
|
||||||
|
|
||||||
|
typedef struct census_log_block_list_struct { |
||||||
|
struct census_log_block_list_struct* next; |
||||||
|
struct census_log_block_list_struct* prev; |
||||||
|
struct census_log_block* block; |
||||||
|
} cl_block_list_struct; |
||||||
|
|
||||||
|
typedef struct census_log_block { |
||||||
|
// Pointer to underlying buffer.
|
||||||
|
char* buffer; |
||||||
|
gpr_atm writer_lock; |
||||||
|
gpr_atm reader_lock; |
||||||
|
// Keeps completely written bytes. Declared atomic because accessed
|
||||||
|
// simultaneously by reader and writer.
|
||||||
|
gpr_atm bytes_committed; |
||||||
|
// Bytes already read.
|
||||||
|
size_t bytes_read; |
||||||
|
// Links for list.
|
||||||
|
cl_block_list_struct link; |
||||||
|
// We want this structure to be cacheline aligned. We assume the following
|
||||||
|
// sizes for the various parts on 32/64bit systems:
|
||||||
|
// type 32b size 64b size
|
||||||
|
// char* 4 8
|
||||||
|
// 3x gpr_atm 12 24
|
||||||
|
// size_t 4 8
|
||||||
|
// cl_block_list_struct 12 24
|
||||||
|
// TOTAL 32 64
|
||||||
|
//
|
||||||
|
// Depending on the size of our cacheline and the architecture, we
|
||||||
|
// selectively add char buffering to this structure. The size is checked
|
||||||
|
// via assert in census_log_initialize().
|
||||||
|
#if defined(GPR_ARCH_64) |
||||||
|
#define CL_BLOCK_PAD_SIZE (GPR_CACHELINE_SIZE - 64) |
||||||
|
#else |
||||||
|
#if defined(GPR_ARCH_32) |
||||||
|
#define CL_BLOCK_PAD_SIZE (GPR_CACHELINE_SIZE - 32) |
||||||
|
#else |
||||||
|
#error "Unknown architecture" |
||||||
|
#endif |
||||||
|
#endif |
||||||
|
#if CL_BLOCK_PAD_SIZE > 0 |
||||||
|
char padding[CL_BLOCK_PAD_SIZE]; |
||||||
|
#endif |
||||||
|
} cl_block; |
||||||
|
|
||||||
|
// A list of cl_blocks, doubly-linked through cl_block::link.
|
||||||
|
typedef struct census_log_block_list { |
||||||
|
int32_t count; // Number of items in list.
|
||||||
|
cl_block_list_struct ht; // head/tail of linked list.
|
||||||
|
} cl_block_list; |
||||||
|
|
||||||
|
// Cacheline aligned block pointers to avoid false sharing. Block pointer must
|
||||||
|
// be initialized via set_block(), before calling other functions
|
||||||
|
typedef struct census_log_core_local_block { |
||||||
|
gpr_atm block; |
||||||
|
// Ensure cachline alignment: we assume sizeof(gpr_atm) == 4 or 8
|
||||||
|
#if defined(GPR_ARCH_64) |
||||||
|
#define CL_CORE_LOCAL_BLOCK_PAD_SIZE (GPR_CACHELINE_SIZE - 8) |
||||||
|
#else |
||||||
|
#if defined(GPR_ARCH_32) |
||||||
|
#define CL_CORE_LOCAL_BLOCK_PAD_SIZE (GPR_CACHELINE_SIZE - 4) |
||||||
|
#else |
||||||
|
#error "Unknown architecture" |
||||||
|
#endif |
||||||
|
#endif |
||||||
|
#if CL_CORE_LOCAL_BLOCK_PAD_SIZE > 0 |
||||||
|
char padding[CL_CORE_LOCAL_BLOCK_PAD_SIZE]; |
||||||
|
#endif |
||||||
|
} cl_core_local_block; |
||||||
|
|
||||||
|
struct census_log { |
||||||
|
int discard_old_records; |
||||||
|
// Number of cores (aka hardware-contexts)
|
||||||
|
unsigned num_cores; |
||||||
|
// number of CENSUS_LOG_2_MAX_RECORD_SIZE blocks in log
|
||||||
|
uint32_t num_blocks; |
||||||
|
cl_block* blocks; // Block metadata.
|
||||||
|
cl_core_local_block* core_local_blocks; // Keeps core to block mappings.
|
||||||
|
gpr_mu lock; |
||||||
|
int initialized; // has log been initialized?
|
||||||
|
// Keeps the state of the reader iterator. A value of 0 indicates that
|
||||||
|
// iterator has reached the end. census_log_init_reader() resets the value
|
||||||
|
// to num_core to restart iteration.
|
||||||
|
uint32_t read_iterator_state; |
||||||
|
// Points to the block being read. If non-NULL, the block is locked for
|
||||||
|
// reading(block_being_read_->reader_lock is held).
|
||||||
|
cl_block* block_being_read; |
||||||
|
char* buffer; |
||||||
|
cl_block_list free_block_list; |
||||||
|
cl_block_list dirty_block_list; |
||||||
|
gpr_atm out_of_space_count; |
||||||
|
}; |
||||||
|
|
||||||
|
// Single internal log.
|
||||||
|
static struct census_log g_log; |
||||||
|
|
||||||
|
// Functions that operate on an atomic memory location used as a lock.
|
||||||
|
|
||||||
|
// Returns non-zero if lock is acquired.
|
||||||
|
static int cl_try_lock(gpr_atm* lock) { return gpr_atm_acq_cas(lock, 0, 1); } |
||||||
|
|
||||||
|
static void cl_unlock(gpr_atm* lock) { gpr_atm_rel_store(lock, 0); } |
||||||
|
|
||||||
|
// Functions that operate on cl_core_local_block's.
|
||||||
|
|
||||||
|
static void cl_core_local_block_set_block(cl_core_local_block* clb, |
||||||
|
cl_block* block) { |
||||||
|
gpr_atm_rel_store(&clb->block, (gpr_atm)block); |
||||||
|
} |
||||||
|
|
||||||
|
static cl_block* cl_core_local_block_get_block(cl_core_local_block* clb) { |
||||||
|
return (cl_block*)gpr_atm_acq_load(&clb->block); |
||||||
|
} |
||||||
|
|
||||||
|
// Functions that operate on cl_block_list_struct's.
|
||||||
|
|
||||||
|
static void cl_block_list_struct_initialize(cl_block_list_struct* bls, |
||||||
|
cl_block* block) { |
||||||
|
bls->next = bls->prev = bls; |
||||||
|
bls->block = block; |
||||||
|
} |
||||||
|
|
||||||
|
// Functions that operate on cl_block_list's.
|
||||||
|
|
||||||
|
static void cl_block_list_initialize(cl_block_list* list) { |
||||||
|
list->count = 0; |
||||||
|
cl_block_list_struct_initialize(&list->ht, NULL); |
||||||
|
} |
||||||
|
|
||||||
|
// Returns head of *this, or NULL if empty.
|
||||||
|
static cl_block* cl_block_list_head(cl_block_list* list) { |
||||||
|
return list->ht.next->block; |
||||||
|
} |
||||||
|
|
||||||
|
// Insert element *e after *pos.
|
||||||
|
static void cl_block_list_insert(cl_block_list* list, cl_block_list_struct* pos, |
||||||
|
cl_block_list_struct* e) { |
||||||
|
list->count++; |
||||||
|
e->next = pos->next; |
||||||
|
e->prev = pos; |
||||||
|
e->next->prev = e; |
||||||
|
e->prev->next = e; |
||||||
|
} |
||||||
|
|
||||||
|
// Insert block at the head of the list
|
||||||
|
static void cl_block_list_insert_at_head(cl_block_list* list, cl_block* block) { |
||||||
|
cl_block_list_insert(list, &list->ht, &block->link); |
||||||
|
} |
||||||
|
|
||||||
|
// Insert block at the tail of the list.
|
||||||
|
static void cl_block_list_insert_at_tail(cl_block_list* list, cl_block* block) { |
||||||
|
cl_block_list_insert(list, list->ht.prev, &block->link); |
||||||
|
} |
||||||
|
|
||||||
|
// Removes block *b. Requires *b be in the list.
|
||||||
|
static void cl_block_list_remove(cl_block_list* list, cl_block* b) { |
||||||
|
list->count--; |
||||||
|
b->link.next->prev = b->link.prev; |
||||||
|
b->link.prev->next = b->link.next; |
||||||
|
} |
||||||
|
|
||||||
|
// Functions that operate on cl_block's
|
||||||
|
|
||||||
|
static void cl_block_initialize(cl_block* block, char* buffer) { |
||||||
|
block->buffer = buffer; |
||||||
|
gpr_atm_rel_store(&block->writer_lock, 0); |
||||||
|
gpr_atm_rel_store(&block->reader_lock, 0); |
||||||
|
gpr_atm_rel_store(&block->bytes_committed, 0); |
||||||
|
block->bytes_read = 0; |
||||||
|
cl_block_list_struct_initialize(&block->link, block); |
||||||
|
} |
||||||
|
|
||||||
|
// Guards against exposing partially written buffer to the reader.
|
||||||
|
static void cl_block_set_bytes_committed(cl_block* block, |
||||||
|
size_t bytes_committed) { |
||||||
|
gpr_atm_rel_store(&block->bytes_committed, (gpr_atm)bytes_committed); |
||||||
|
} |
||||||
|
|
||||||
|
static size_t cl_block_get_bytes_committed(cl_block* block) { |
||||||
|
return (size_t)gpr_atm_acq_load(&block->bytes_committed); |
||||||
|
} |
||||||
|
|
||||||
|
// Tries to disable future read/write access to this block. Succeeds if:
|
||||||
|
// - no in-progress write AND
|
||||||
|
// - no in-progress read AND
|
||||||
|
// - 'discard_data' set to true OR no unread data
|
||||||
|
// On success, clears the block state and returns with writer_lock_ and
|
||||||
|
// reader_lock_ held. These locks are released by a subsequent
|
||||||
|
// cl_block_access_enable() call.
|
||||||
|
static bool cl_block_try_disable_access(cl_block* block, int discard_data) { |
||||||
|
if (!cl_try_lock(&block->writer_lock)) { |
||||||
|
return false; |
||||||
|
} |
||||||
|
if (!cl_try_lock(&block->reader_lock)) { |
||||||
|
cl_unlock(&block->writer_lock); |
||||||
|
return false; |
||||||
|
} |
||||||
|
if (!discard_data && |
||||||
|
(block->bytes_read != cl_block_get_bytes_committed(block))) { |
||||||
|
cl_unlock(&block->reader_lock); |
||||||
|
cl_unlock(&block->writer_lock); |
||||||
|
return false; |
||||||
|
} |
||||||
|
cl_block_set_bytes_committed(block, 0); |
||||||
|
block->bytes_read = 0; |
||||||
|
return true; |
||||||
|
} |
||||||
|
|
||||||
|
static void cl_block_enable_access(cl_block* block) { |
||||||
|
cl_unlock(&block->reader_lock); |
||||||
|
cl_unlock(&block->writer_lock); |
||||||
|
} |
||||||
|
|
||||||
|
// Returns with writer_lock held.
|
||||||
|
static void* cl_block_start_write(cl_block* block, size_t size) { |
||||||
|
if (!cl_try_lock(&block->writer_lock)) { |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
size_t bytes_committed = cl_block_get_bytes_committed(block); |
||||||
|
if (bytes_committed + size > CENSUS_LOG_MAX_RECORD_SIZE) { |
||||||
|
cl_unlock(&block->writer_lock); |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
return block->buffer + bytes_committed; |
||||||
|
} |
||||||
|
|
||||||
|
// Releases writer_lock and increments committed bytes by 'bytes_written'.
|
||||||
|
// 'bytes_written' must be <= 'size' specified in the corresponding
|
||||||
|
// StartWrite() call. This function is thread-safe.
|
||||||
|
static void cl_block_end_write(cl_block* block, size_t bytes_written) { |
||||||
|
cl_block_set_bytes_committed( |
||||||
|
block, cl_block_get_bytes_committed(block) + bytes_written); |
||||||
|
cl_unlock(&block->writer_lock); |
||||||
|
} |
||||||
|
|
||||||
|
// Returns a pointer to the first unread byte in buffer. The number of bytes
|
||||||
|
// available are returned in 'bytes_available'. Acquires reader lock that is
|
||||||
|
// released by a subsequent cl_block_end_read() call. Returns NULL if:
|
||||||
|
// - read in progress
|
||||||
|
// - no data available
|
||||||
|
static void* cl_block_start_read(cl_block* block, size_t* bytes_available) { |
||||||
|
if (!cl_try_lock(&block->reader_lock)) { |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
// bytes_committed may change from under us. Use bytes_available to update
|
||||||
|
// bytes_read below.
|
||||||
|
size_t bytes_committed = cl_block_get_bytes_committed(block); |
||||||
|
GPR_ASSERT(bytes_committed >= block->bytes_read); |
||||||
|
*bytes_available = bytes_committed - block->bytes_read; |
||||||
|
if (*bytes_available == 0) { |
||||||
|
cl_unlock(&block->reader_lock); |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
void* record = block->buffer + block->bytes_read; |
||||||
|
block->bytes_read += *bytes_available; |
||||||
|
return record; |
||||||
|
} |
||||||
|
|
||||||
|
static void cl_block_end_read(cl_block* block) { |
||||||
|
cl_unlock(&block->reader_lock); |
||||||
|
} |
||||||
|
|
||||||
|
// Internal functions operating on g_log
|
||||||
|
|
||||||
|
// Allocates a new free block (or recycles an available dirty block if log is
|
||||||
|
// configured to discard old records). Returns NULL if out-of-space.
|
||||||
|
static cl_block* cl_allocate_block(void) { |
||||||
|
cl_block* block = cl_block_list_head(&g_log.free_block_list); |
||||||
|
if (block != NULL) { |
||||||
|
cl_block_list_remove(&g_log.free_block_list, block); |
||||||
|
return block; |
||||||
|
} |
||||||
|
if (!g_log.discard_old_records) { |
||||||
|
// No free block and log is configured to keep old records.
|
||||||
|
return NULL; |
||||||
|
} |
||||||
|
// Recycle dirty block. Start from the oldest.
|
||||||
|
for (block = cl_block_list_head(&g_log.dirty_block_list); block != NULL; |
||||||
|
block = block->link.next->block) { |
||||||
|
if (cl_block_try_disable_access(block, 1 /* discard data */)) { |
||||||
|
cl_block_list_remove(&g_log.dirty_block_list, block); |
||||||
|
return block; |
||||||
|
} |
||||||
|
} |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
|
||||||
|
// Allocates a new block and updates core id => block mapping. 'old_block'
|
||||||
|
// points to the block that the caller thinks is attached to
|
||||||
|
// 'core_id'. 'old_block' may be NULL. Returns true if:
|
||||||
|
// - allocated a new block OR
|
||||||
|
// - 'core_id' => 'old_block' mapping changed (another thread allocated a
|
||||||
|
// block before lock was acquired).
|
||||||
|
static bool cl_allocate_core_local_block(uint32_t core_id, |
||||||
|
cl_block* old_block) { |
||||||
|
// Now that we have the lock, check if core-local mapping has changed.
|
||||||
|
cl_core_local_block* core_local_block = &g_log.core_local_blocks[core_id]; |
||||||
|
cl_block* block = cl_core_local_block_get_block(core_local_block); |
||||||
|
if ((block != NULL) && (block != old_block)) { |
||||||
|
return true; |
||||||
|
} |
||||||
|
if (block != NULL) { |
||||||
|
cl_core_local_block_set_block(core_local_block, NULL); |
||||||
|
cl_block_list_insert_at_tail(&g_log.dirty_block_list, block); |
||||||
|
} |
||||||
|
block = cl_allocate_block(); |
||||||
|
if (block == NULL) { |
||||||
|
return false; |
||||||
|
} |
||||||
|
cl_core_local_block_set_block(core_local_block, block); |
||||||
|
cl_block_enable_access(block); |
||||||
|
return true; |
||||||
|
} |
||||||
|
|
||||||
|
static cl_block* cl_get_block(void* record) { |
||||||
|
uintptr_t p = (uintptr_t)((char*)record - g_log.buffer); |
||||||
|
uintptr_t index = p >> CENSUS_LOG_2_MAX_RECORD_SIZE; |
||||||
|
return &g_log.blocks[index]; |
||||||
|
} |
||||||
|
|
||||||
|
// Gets the next block to read and tries to free 'prev' block (if not NULL).
|
||||||
|
// Returns NULL if reached the end.
|
||||||
|
static cl_block* cl_next_block_to_read(cl_block* prev) { |
||||||
|
cl_block* block = NULL; |
||||||
|
if (g_log.read_iterator_state == g_log.num_cores) { |
||||||
|
// We are traversing dirty list; find the next dirty block.
|
||||||
|
if (prev != NULL) { |
||||||
|
// Try to free the previous block if there is no unread data. This
|
||||||
|
// block
|
||||||
|
// may have unread data if previously incomplete record completed
|
||||||
|
// between
|
||||||
|
// read_next() calls.
|
||||||
|
block = prev->link.next->block; |
||||||
|
if (cl_block_try_disable_access(prev, 0 /* do not discard data */)) { |
||||||
|
cl_block_list_remove(&g_log.dirty_block_list, prev); |
||||||
|
cl_block_list_insert_at_head(&g_log.free_block_list, prev); |
||||||
|
} |
||||||
|
} else { |
||||||
|
block = cl_block_list_head(&g_log.dirty_block_list); |
||||||
|
} |
||||||
|
if (block != NULL) { |
||||||
|
return block; |
||||||
|
} |
||||||
|
// We are done with the dirty list; moving on to core-local blocks.
|
||||||
|
} |
||||||
|
while (g_log.read_iterator_state > 0) { |
||||||
|
g_log.read_iterator_state--; |
||||||
|
block = cl_core_local_block_get_block( |
||||||
|
&g_log.core_local_blocks[g_log.read_iterator_state]); |
||||||
|
if (block != NULL) { |
||||||
|
return block; |
||||||
|
} |
||||||
|
} |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
|
||||||
|
#define CL_LOG_2_MB 20 // 2^20 = 1MB
|
||||||
|
|
||||||
|
// External functions: primary stats_log interface
|
||||||
|
void census_log_initialize(size_t size_in_mb, int discard_old_records) { |
||||||
|
// Check cacheline alignment.
|
||||||
|
GPR_ASSERT(sizeof(cl_block) % GPR_CACHELINE_SIZE == 0); |
||||||
|
GPR_ASSERT(sizeof(cl_core_local_block) % GPR_CACHELINE_SIZE == 0); |
||||||
|
GPR_ASSERT(!g_log.initialized); |
||||||
|
g_log.discard_old_records = discard_old_records; |
||||||
|
g_log.num_cores = gpr_cpu_num_cores(); |
||||||
|
// Ensure that we will not get any overflow in calaculating num_blocks
|
||||||
|
GPR_ASSERT(CL_LOG_2_MB >= CENSUS_LOG_2_MAX_RECORD_SIZE); |
||||||
|
GPR_ASSERT(size_in_mb < 1000); |
||||||
|
// Ensure at least 2x as many blocks as there are cores.
|
||||||
|
g_log.num_blocks = |
||||||
|
(uint32_t)GPR_MAX(2 * g_log.num_cores, (size_in_mb << CL_LOG_2_MB) >> |
||||||
|
CENSUS_LOG_2_MAX_RECORD_SIZE); |
||||||
|
gpr_mu_init(&g_log.lock); |
||||||
|
g_log.read_iterator_state = 0; |
||||||
|
g_log.block_being_read = NULL; |
||||||
|
g_log.core_local_blocks = (cl_core_local_block*)gpr_malloc_aligned( |
||||||
|
g_log.num_cores * sizeof(cl_core_local_block), GPR_CACHELINE_SIZE_LOG); |
||||||
|
memset(g_log.core_local_blocks, 0, |
||||||
|
g_log.num_cores * sizeof(cl_core_local_block)); |
||||||
|
g_log.blocks = (cl_block*)gpr_malloc_aligned( |
||||||
|
g_log.num_blocks * sizeof(cl_block), GPR_CACHELINE_SIZE_LOG); |
||||||
|
memset(g_log.blocks, 0, g_log.num_blocks * sizeof(cl_block)); |
||||||
|
g_log.buffer = gpr_malloc(g_log.num_blocks * CENSUS_LOG_MAX_RECORD_SIZE); |
||||||
|
memset(g_log.buffer, 0, g_log.num_blocks * CENSUS_LOG_MAX_RECORD_SIZE); |
||||||
|
cl_block_list_initialize(&g_log.free_block_list); |
||||||
|
cl_block_list_initialize(&g_log.dirty_block_list); |
||||||
|
for (uint32_t i = 0; i < g_log.num_blocks; ++i) { |
||||||
|
cl_block* block = g_log.blocks + i; |
||||||
|
cl_block_initialize(block, g_log.buffer + (CENSUS_LOG_MAX_RECORD_SIZE * i)); |
||||||
|
cl_block_try_disable_access(block, 1 /* discard data */); |
||||||
|
cl_block_list_insert_at_tail(&g_log.free_block_list, block); |
||||||
|
} |
||||||
|
gpr_atm_rel_store(&g_log.out_of_space_count, 0); |
||||||
|
g_log.initialized = 1; |
||||||
|
} |
||||||
|
|
||||||
|
void census_log_shutdown(void) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
gpr_mu_destroy(&g_log.lock); |
||||||
|
gpr_free_aligned(g_log.core_local_blocks); |
||||||
|
g_log.core_local_blocks = NULL; |
||||||
|
gpr_free_aligned(g_log.blocks); |
||||||
|
g_log.blocks = NULL; |
||||||
|
gpr_free(g_log.buffer); |
||||||
|
g_log.buffer = NULL; |
||||||
|
g_log.initialized = 0; |
||||||
|
} |
||||||
|
|
||||||
|
void* census_log_start_write(size_t size) { |
||||||
|
// Used to bound number of times block allocation is attempted.
|
||||||
|
GPR_ASSERT(size > 0); |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
if (size > CENSUS_LOG_MAX_RECORD_SIZE) { |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
uint32_t attempts_remaining = g_log.num_blocks; |
||||||
|
uint32_t core_id = gpr_cpu_current_cpu(); |
||||||
|
do { |
||||||
|
void* record = NULL; |
||||||
|
cl_block* block = |
||||||
|
cl_core_local_block_get_block(&g_log.core_local_blocks[core_id]); |
||||||
|
if (block && (record = cl_block_start_write(block, size))) { |
||||||
|
return record; |
||||||
|
} |
||||||
|
// Need to allocate a new block. We are here if:
|
||||||
|
// - No block associated with the core OR
|
||||||
|
// - Write in-progress on the block OR
|
||||||
|
// - block is out of space
|
||||||
|
gpr_mu_lock(&g_log.lock); |
||||||
|
bool allocated = cl_allocate_core_local_block(core_id, block); |
||||||
|
gpr_mu_unlock(&g_log.lock); |
||||||
|
if (!allocated) { |
||||||
|
gpr_atm_no_barrier_fetch_add(&g_log.out_of_space_count, 1); |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
} while (attempts_remaining--); |
||||||
|
// Give up.
|
||||||
|
gpr_atm_no_barrier_fetch_add(&g_log.out_of_space_count, 1); |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
|
||||||
|
void census_log_end_write(void* record, size_t bytes_written) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
cl_block_end_write(cl_get_block(record), bytes_written); |
||||||
|
} |
||||||
|
|
||||||
|
void census_log_init_reader(void) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
gpr_mu_lock(&g_log.lock); |
||||||
|
// If a block is locked for reading unlock it.
|
||||||
|
if (g_log.block_being_read != NULL) { |
||||||
|
cl_block_end_read(g_log.block_being_read); |
||||||
|
g_log.block_being_read = NULL; |
||||||
|
} |
||||||
|
g_log.read_iterator_state = g_log.num_cores; |
||||||
|
gpr_mu_unlock(&g_log.lock); |
||||||
|
} |
||||||
|
|
||||||
|
const void* census_log_read_next(size_t* bytes_available) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
gpr_mu_lock(&g_log.lock); |
||||||
|
if (g_log.block_being_read != NULL) { |
||||||
|
cl_block_end_read(g_log.block_being_read); |
||||||
|
} |
||||||
|
do { |
||||||
|
g_log.block_being_read = cl_next_block_to_read(g_log.block_being_read); |
||||||
|
if (g_log.block_being_read != NULL) { |
||||||
|
void* record = |
||||||
|
cl_block_start_read(g_log.block_being_read, bytes_available); |
||||||
|
if (record != NULL) { |
||||||
|
gpr_mu_unlock(&g_log.lock); |
||||||
|
return record; |
||||||
|
} |
||||||
|
} |
||||||
|
} while (g_log.block_being_read != NULL); |
||||||
|
gpr_mu_unlock(&g_log.lock); |
||||||
|
return NULL; |
||||||
|
} |
||||||
|
|
||||||
|
size_t census_log_remaining_space(void) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
size_t space = 0; |
||||||
|
gpr_mu_lock(&g_log.lock); |
||||||
|
if (g_log.discard_old_records) { |
||||||
|
// Remaining space is not meaningful; just return the entire log space.
|
||||||
|
space = g_log.num_blocks << CENSUS_LOG_2_MAX_RECORD_SIZE; |
||||||
|
} else { |
||||||
|
GPR_ASSERT(g_log.free_block_list.count >= 0); |
||||||
|
space = (size_t)g_log.free_block_list.count * CENSUS_LOG_MAX_RECORD_SIZE; |
||||||
|
} |
||||||
|
gpr_mu_unlock(&g_log.lock); |
||||||
|
return space; |
||||||
|
} |
||||||
|
|
||||||
|
int64_t census_log_out_of_space_count(void) { |
||||||
|
GPR_ASSERT(g_log.initialized); |
||||||
|
return gpr_atm_acq_load(&g_log.out_of_space_count); |
||||||
|
} |
@ -0,0 +1,95 @@ |
|||||||
|
/*
|
||||||
|
* |
||||||
|
* Copyright 2015-2016, Google Inc. |
||||||
|
* All rights reserved. |
||||||
|
* |
||||||
|
* Redistribution and use in source and binary forms, with or without |
||||||
|
* modification, are permitted provided that the following conditions are |
||||||
|
* met: |
||||||
|
* |
||||||
|
* * Redistributions of source code must retain the above copyright |
||||||
|
* notice, this list of conditions and the following disclaimer. |
||||||
|
* * Redistributions in binary form must reproduce the above |
||||||
|
* copyright notice, this list of conditions and the following disclaimer |
||||||
|
* in the documentation and/or other materials provided with the |
||||||
|
* distribution. |
||||||
|
* * Neither the name of Google Inc. nor the names of its |
||||||
|
* contributors may be used to endorse or promote products derived from |
||||||
|
* this software without specific prior written permission. |
||||||
|
* |
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
||||||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
||||||
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
||||||
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
||||||
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||||||
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||||||
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||||||
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||||||
|
* |
||||||
|
*/ |
||||||
|
|
||||||
|
/* A very fast in-memory log, optimized for multiple writers. */ |
||||||
|
|
||||||
|
#ifndef GRPC_INTERNAL_CORE_CENSUS_MLOG_H |
||||||
|
#define GRPC_INTERNAL_CORE_CENSUS_MLOG_H |
||||||
|
|
||||||
|
#include <grpc/support/port_platform.h> |
||||||
|
#include <stddef.h> |
||||||
|
|
||||||
|
/* Maximum record size, in bytes. */ |
||||||
|
#define CENSUS_LOG_2_MAX_RECORD_SIZE 14 /* 2^14 = 16KB */ |
||||||
|
#define CENSUS_LOG_MAX_RECORD_SIZE (1 << CENSUS_LOG_2_MAX_RECORD_SIZE) |
||||||
|
|
||||||
|
/* Initialize the statistics logging subsystem with the given log size. A log
|
||||||
|
size of 0 will result in the smallest possible log for the platform |
||||||
|
(approximately CENSUS_LOG_MAX_RECORD_SIZE * gpr_cpu_num_cores()). If |
||||||
|
discard_old_records is non-zero, then new records will displace older ones |
||||||
|
when the log is full. This function must be called before any other |
||||||
|
census_log functions. |
||||||
|
*/ |
||||||
|
void census_log_initialize(size_t size_in_mb, int discard_old_records); |
||||||
|
|
||||||
|
/* Shutdown the logging subsystem. Caller must ensure that:
|
||||||
|
- no in progress or future call to any census_log functions |
||||||
|
- no incomplete records |
||||||
|
*/ |
||||||
|
void census_log_shutdown(void); |
||||||
|
|
||||||
|
/* Allocates and returns a 'size' bytes record and marks it in use. A
|
||||||
|
subsequent census_log_end_write() marks the record complete. The |
||||||
|
'bytes_written' census_log_end_write() argument must be <= |
||||||
|
'size'. Returns NULL if out-of-space AND: |
||||||
|
- log is configured to keep old records OR |
||||||
|
- all blocks are pinned by incomplete records. |
||||||
|
*/ |
||||||
|
void* census_log_start_write(size_t size); |
||||||
|
|
||||||
|
void census_log_end_write(void* record, size_t bytes_written); |
||||||
|
|
||||||
|
void census_log_init_reader(void); |
||||||
|
|
||||||
|
/* census_log_read_next() iterates over blocks with data and for each block
|
||||||
|
returns a pointer to the first unread byte. The number of bytes that can be |
||||||
|
read are returned in 'bytes_available'. Reader is expected to read all |
||||||
|
available data. Reading the data consumes it i.e. it cannot be read again. |
||||||
|
census_log_read_next() returns NULL if the end is reached i.e last block |
||||||
|
is read. census_log_init_reader() starts the iteration or aborts the |
||||||
|
current iteration. |
||||||
|
*/ |
||||||
|
const void* census_log_read_next(size_t* bytes_available); |
||||||
|
|
||||||
|
/* Returns estimated remaining space across all blocks, in bytes. If log is
|
||||||
|
configured to discard old records, returns total log space. Otherwise, |
||||||
|
returns space available in empty blocks (partially filled blocks are |
||||||
|
treated as full). |
||||||
|
*/ |
||||||
|
size_t census_log_remaining_space(void); |
||||||
|
|
||||||
|
/* Returns the number of times gprc_stats_log_start_write() failed due to
|
||||||
|
out-of-space. */ |
||||||
|
int64_t census_log_out_of_space_count(void); |
||||||
|
|
||||||
|
#endif /* GRPC_INTERNAL_CORE_CENSUS_LOG_H */ |
@ -0,0 +1,259 @@ |
|||||||
|
//
|
||||||
|
//
|
||||||
|
// Copyright 2016, Google Inc.
|
||||||
|
// All rights reserved.
|
||||||
|
//
|
||||||
|
// Redistribution and use in source and binary forms, with or without
|
||||||
|
// modification, are permitted provided that the following conditions are
|
||||||
|
// met:
|
||||||
|
//
|
||||||
|
// * Redistributions of source code must retain the above copyright
|
||||||
|
// notice, this list of conditions and the following disclaimer.
|
||||||
|
// * Redistributions in binary form must reproduce the above
|
||||||
|
// copyright notice, this list of conditions and the following disclaimer
|
||||||
|
// in the documentation and/or other materials provided with the
|
||||||
|
// distribution.
|
||||||
|
// * Neither the name of Google Inc. nor the names of its
|
||||||
|
// contributors may be used to endorse or promote products derived from
|
||||||
|
// this software without specific prior written permission.
|
||||||
|
//
|
||||||
|
// THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
|
||||||
|
// "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
|
||||||
|
// LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
|
||||||
|
// A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
|
||||||
|
// OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
|
||||||
|
// SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
|
||||||
|
// LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
|
||||||
|
// DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
|
||||||
|
// THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
|
||||||
|
// (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
|
||||||
|
// OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
||||||
|
//
|
||||||
|
//
|
||||||
|
|
||||||
|
#include "src/core/client_config/subchannel_index.h" |
||||||
|
|
||||||
|
#include <stdbool.h> |
||||||
|
#include <string.h> |
||||||
|
|
||||||
|
#include <grpc/support/alloc.h> |
||||||
|
#include <grpc/support/avl.h> |
||||||
|
#include <grpc/support/tls.h> |
||||||
|
|
||||||
|
#include "src/core/channel/channel_args.h" |
||||||
|
|
||||||
|
// a map of subchannel_key --> subchannel, used for detecting connections
|
||||||
|
// to the same destination in order to share them
|
||||||
|
static gpr_avl g_subchannel_index; |
||||||
|
|
||||||
|
static gpr_mu g_mu; |
||||||
|
|
||||||
|
struct grpc_subchannel_key { |
||||||
|
grpc_connector *connector; |
||||||
|
grpc_subchannel_args args; |
||||||
|
}; |
||||||
|
|
||||||
|
GPR_TLS_DECL(subchannel_index_exec_ctx); |
||||||
|
|
||||||
|
static void enter_ctx(grpc_exec_ctx *exec_ctx) { |
||||||
|
GPR_ASSERT(gpr_tls_get(&subchannel_index_exec_ctx) == 0); |
||||||
|
gpr_tls_set(&subchannel_index_exec_ctx, (intptr_t)exec_ctx); |
||||||
|
} |
||||||
|
|
||||||
|
static void leave_ctx(grpc_exec_ctx *exec_ctx) { |
||||||
|
GPR_ASSERT(gpr_tls_get(&subchannel_index_exec_ctx) == (intptr_t)exec_ctx); |
||||||
|
gpr_tls_set(&subchannel_index_exec_ctx, 0); |
||||||
|
} |
||||||
|
|
||||||
|
static grpc_exec_ctx *current_ctx() { |
||||||
|
grpc_exec_ctx *c = (grpc_exec_ctx *)gpr_tls_get(&subchannel_index_exec_ctx); |
||||||
|
GPR_ASSERT(c != NULL); |
||||||
|
return c; |
||||||
|
} |
||||||
|
|
||||||
|
static grpc_subchannel_key *create_key( |
||||||
|
grpc_connector *connector, grpc_subchannel_args *args, |
||||||
|
grpc_channel_args *(*copy_channel_args)(const grpc_channel_args *args)) { |
||||||
|
grpc_subchannel_key *k = gpr_malloc(sizeof(*k)); |
||||||
|
k->connector = grpc_connector_ref(connector); |
||||||
|
k->args.filter_count = args->filter_count; |
||||||
|
k->args.filters = gpr_malloc(sizeof(*k->args.filters) * k->args.filter_count); |
||||||
|
memcpy((grpc_channel_filter *)k->args.filters, args->filters, |
||||||
|
sizeof(*k->args.filters) * k->args.filter_count); |
||||||
|
k->args.addr_len = args->addr_len; |
||||||
|
k->args.addr = gpr_malloc(args->addr_len); |
||||||
|
memcpy(k->args.addr, args->addr, k->args.addr_len); |
||||||
|
k->args.args = copy_channel_args(args->args); |
||||||
|
return k; |
||||||
|
} |
||||||
|
|
||||||
|
grpc_subchannel_key *grpc_subchannel_key_create(grpc_connector *connector, |
||||||
|
grpc_subchannel_args *args) { |
||||||
|
return create_key(connector, args, grpc_channel_args_normalize); |
||||||
|
} |
||||||
|
|
||||||
|
static grpc_subchannel_key *subchannel_key_copy(grpc_subchannel_key *k) { |
||||||
|
return create_key(k->connector, &k->args, grpc_channel_args_copy); |
||||||
|
} |
||||||
|
|
||||||
|
static int subchannel_key_compare(grpc_subchannel_key *a, |
||||||
|
grpc_subchannel_key *b) { |
||||||
|
int c = GPR_ICMP(a->connector, b->connector); |
||||||
|
if (c != 0) return c; |
||||||
|
c = GPR_ICMP(a->args.addr_len, b->args.addr_len); |
||||||
|
if (c != 0) return c; |
||||||
|
c = GPR_ICMP(a->args.filter_count, b->args.filter_count); |
||||||
|
if (c != 0) return c; |
||||||
|
c = memcmp(a->args.addr, b->args.addr, a->args.addr_len); |
||||||
|
if (c != 0) return c; |
||||||
|
c = memcmp(a->args.filters, b->args.filters, |
||||||
|
a->args.filter_count * sizeof(*a->args.filters)); |
||||||
|
return grpc_channel_args_compare(a->args.args, b->args.args); |
||||||
|
} |
||||||
|
|
||||||
|
void grpc_subchannel_key_destroy(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *k) { |
||||||
|
grpc_connector_unref(exec_ctx, k->connector); |
||||||
|
gpr_free(k->args.addr); |
||||||
|
gpr_free((grpc_channel_args *)k->args.filters); |
||||||
|
grpc_channel_args_destroy((grpc_channel_args *)k->args.args); |
||||||
|
gpr_free(k); |
||||||
|
} |
||||||
|
|
||||||
|
static void sck_avl_destroy(void *p) { |
||||||
|
grpc_subchannel_key_destroy(current_ctx(), p); |
||||||
|
} |
||||||
|
|
||||||
|
static void *sck_avl_copy(void *p) { return subchannel_key_copy(p); } |
||||||
|
|
||||||
|
static long sck_avl_compare(void *a, void *b) { |
||||||
|
return subchannel_key_compare(a, b); |
||||||
|
} |
||||||
|
|
||||||
|
static void scv_avl_destroy(void *p) { |
||||||
|
GRPC_SUBCHANNEL_WEAK_UNREF(current_ctx(), p, "subchannel_index"); |
||||||
|
} |
||||||
|
|
||||||
|
static void *scv_avl_copy(void *p) { |
||||||
|
GRPC_SUBCHANNEL_WEAK_REF(p, "subchannel_index"); |
||||||
|
return p; |
||||||
|
} |
||||||
|
|
||||||
|
static const gpr_avl_vtable subchannel_avl_vtable = { |
||||||
|
.destroy_key = sck_avl_destroy, |
||||||
|
.copy_key = sck_avl_copy, |
||||||
|
.compare_keys = sck_avl_compare, |
||||||
|
.destroy_value = scv_avl_destroy, |
||||||
|
.copy_value = scv_avl_copy}; |
||||||
|
|
||||||
|
void grpc_subchannel_index_init(void) { |
||||||
|
g_subchannel_index = gpr_avl_create(&subchannel_avl_vtable); |
||||||
|
gpr_mu_init(&g_mu); |
||||||
|
} |
||||||
|
|
||||||
|
void grpc_subchannel_index_shutdown(void) { |
||||||
|
gpr_mu_destroy(&g_mu); |
||||||
|
gpr_avl_unref(g_subchannel_index); |
||||||
|
} |
||||||
|
|
||||||
|
grpc_subchannel *grpc_subchannel_index_find(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key) { |
||||||
|
enter_ctx(exec_ctx); |
||||||
|
|
||||||
|
// Lock, and take a reference to the subchannel index.
|
||||||
|
// We don't need to do the search under a lock as avl's are immutable.
|
||||||
|
gpr_mu_lock(&g_mu); |
||||||
|
gpr_avl index = gpr_avl_ref(g_subchannel_index); |
||||||
|
gpr_mu_unlock(&g_mu); |
||||||
|
|
||||||
|
grpc_subchannel *c = |
||||||
|
GRPC_SUBCHANNEL_REF_FROM_WEAK_REF(gpr_avl_get(index, key), "index_find"); |
||||||
|
gpr_avl_unref(index); |
||||||
|
|
||||||
|
leave_ctx(exec_ctx); |
||||||
|
return c; |
||||||
|
} |
||||||
|
|
||||||
|
grpc_subchannel *grpc_subchannel_index_register(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key, |
||||||
|
grpc_subchannel *constructed) { |
||||||
|
enter_ctx(exec_ctx); |
||||||
|
|
||||||
|
grpc_subchannel *c = NULL; |
||||||
|
|
||||||
|
while (c == NULL) { |
||||||
|
// Compare and swap loop:
|
||||||
|
// - take a reference to the current index
|
||||||
|
gpr_mu_lock(&g_mu); |
||||||
|
gpr_avl index = gpr_avl_ref(g_subchannel_index); |
||||||
|
gpr_mu_unlock(&g_mu); |
||||||
|
|
||||||
|
// - Check to see if a subchannel already exists
|
||||||
|
c = gpr_avl_get(index, key); |
||||||
|
if (c != NULL) { |
||||||
|
// yes -> we're done
|
||||||
|
GRPC_SUBCHANNEL_WEAK_UNREF(exec_ctx, constructed, "index_register"); |
||||||
|
} else { |
||||||
|
// no -> update the avl and compare/swap
|
||||||
|
gpr_avl updated = |
||||||
|
gpr_avl_add(gpr_avl_ref(index), subchannel_key_copy(key), |
||||||
|
GRPC_SUBCHANNEL_WEAK_REF(constructed, "index_register")); |
||||||
|
|
||||||
|
// it may happen (but it's expected to be unlikely)
|
||||||
|
// that some other thread has changed the index:
|
||||||
|
// compare/swap here to check that, and retry as necessary
|
||||||
|
gpr_mu_lock(&g_mu); |
||||||
|
if (index.root == g_subchannel_index.root) { |
||||||
|
GPR_SWAP(gpr_avl, updated, g_subchannel_index); |
||||||
|
c = constructed; |
||||||
|
} |
||||||
|
gpr_mu_unlock(&g_mu); |
||||||
|
|
||||||
|
gpr_avl_unref(updated); |
||||||
|
} |
||||||
|
gpr_avl_unref(index); |
||||||
|
} |
||||||
|
|
||||||
|
leave_ctx(exec_ctx); |
||||||
|
|
||||||
|
return c; |
||||||
|
} |
||||||
|
|
||||||
|
void grpc_subchannel_index_unregister(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key, |
||||||
|
grpc_subchannel *constructed) { |
||||||
|
enter_ctx(exec_ctx); |
||||||
|
|
||||||
|
bool done = false; |
||||||
|
while (!done) { |
||||||
|
// Compare and swap loop:
|
||||||
|
// - take a reference to the current index
|
||||||
|
gpr_mu_lock(&g_mu); |
||||||
|
gpr_avl index = gpr_avl_ref(g_subchannel_index); |
||||||
|
gpr_mu_unlock(&g_mu); |
||||||
|
|
||||||
|
// Check to see if this key still refers to the previously
|
||||||
|
// registered subchannel
|
||||||
|
grpc_subchannel *c = gpr_avl_get(index, key); |
||||||
|
if (c != constructed) { |
||||||
|
gpr_avl_unref(index); |
||||||
|
break; |
||||||
|
} |
||||||
|
|
||||||
|
// compare and swap the update (some other thread may have
|
||||||
|
// mutated the index behind us)
|
||||||
|
gpr_avl updated = gpr_avl_remove(gpr_avl_ref(index), key); |
||||||
|
|
||||||
|
gpr_mu_lock(&g_mu); |
||||||
|
if (index.root == g_subchannel_index.root) { |
||||||
|
GPR_SWAP(gpr_avl, updated, g_subchannel_index); |
||||||
|
done = true; |
||||||
|
} |
||||||
|
gpr_mu_unlock(&g_mu); |
||||||
|
|
||||||
|
gpr_avl_unref(updated); |
||||||
|
gpr_avl_unref(index); |
||||||
|
} |
||||||
|
|
||||||
|
leave_ctx(exec_ctx); |
||||||
|
} |
@ -0,0 +1,77 @@ |
|||||||
|
/*
|
||||||
|
* |
||||||
|
* Copyright 2016, Google Inc. |
||||||
|
* All rights reserved. |
||||||
|
* |
||||||
|
* Redistribution and use in source and binary forms, with or without |
||||||
|
* modification, are permitted provided that the following conditions are |
||||||
|
* met: |
||||||
|
* |
||||||
|
* * Redistributions of source code must retain the above copyright |
||||||
|
* notice, this list of conditions and the following disclaimer. |
||||||
|
* * Redistributions in binary form must reproduce the above |
||||||
|
* copyright notice, this list of conditions and the following disclaimer |
||||||
|
* in the documentation and/or other materials provided with the |
||||||
|
* distribution. |
||||||
|
* * Neither the name of Google Inc. nor the names of its |
||||||
|
* contributors may be used to endorse or promote products derived from |
||||||
|
* this software without specific prior written permission. |
||||||
|
* |
||||||
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
||||||
|
* "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
||||||
|
* A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
||||||
|
* OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
||||||
|
* SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
||||||
|
* LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||||||
|
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||||||
|
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||||||
|
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
||||||
|
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||||||
|
* |
||||||
|
*/ |
||||||
|
|
||||||
|
#ifndef GRPC_INTERNAL_CORE_CLIENT_CONFIG_SUBCHANNEL_INDEX_H |
||||||
|
#define GRPC_INTERNAL_CORE_CLIENT_CONFIG_SUBCHANNEL_INDEX_H |
||||||
|
|
||||||
|
#include "src/core/client_config/connector.h" |
||||||
|
#include "src/core/client_config/subchannel.h" |
||||||
|
|
||||||
|
/** \file Provides an index of active subchannels so that they can be
|
||||||
|
shared amongst channels */ |
||||||
|
|
||||||
|
typedef struct grpc_subchannel_key grpc_subchannel_key; |
||||||
|
|
||||||
|
/** Create a key that can be used to uniquely identify a subchannel */ |
||||||
|
grpc_subchannel_key *grpc_subchannel_key_create(grpc_connector *con, |
||||||
|
grpc_subchannel_args *args); |
||||||
|
|
||||||
|
/** Destroy a subchannel key */ |
||||||
|
void grpc_subchannel_key_destroy(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key); |
||||||
|
|
||||||
|
/** Given a subchannel key, find the subchannel registered for it.
|
||||||
|
Returns NULL if no such channel exists. |
||||||
|
Thread-safe. */ |
||||||
|
grpc_subchannel *grpc_subchannel_index_find(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key); |
||||||
|
|
||||||
|
/** Register a subchannel against a key.
|
||||||
|
Takes ownership of \a constructed. |
||||||
|
Returns the registered subchannel. This may be different from |
||||||
|
\a constructed in the case of a registration race. */ |
||||||
|
grpc_subchannel *grpc_subchannel_index_register(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key, |
||||||
|
grpc_subchannel *constructed); |
||||||
|
|
||||||
|
/** Remove \a constructed as the registered subchannel for \a key. */ |
||||||
|
void grpc_subchannel_index_unregister(grpc_exec_ctx *exec_ctx, |
||||||
|
grpc_subchannel_key *key, |
||||||
|
grpc_subchannel *constructed); |
||||||
|
|
||||||
|
/** Initialize the subchannel index (global) */ |
||||||
|
void grpc_subchannel_index_init(void); |
||||||
|
/** Shutdown the subchannel index (global) */ |
||||||
|
void grpc_subchannel_index_shutdown(void); |
||||||
|
|
||||||
|
#endif /* GRPC_INTERNAL_CORE_CLIENT_CONFIG_SUBCHANNEL_INDEX_H */ |
Some files were not shown because too many files have changed in this diff Show More
Loading…
Reference in new issue