Draft EDS API. (#11)
The idea here is to get into the repository something with reasonable fidelity to the early drafts that have been floatedpull/19/head
parent
b8cdf40652
commit
1f6bfe61f1
3 changed files with 214 additions and 0 deletions
@ -0,0 +1,189 @@ |
|||||||
|
syntax = "proto3"; |
||||||
|
|
||||||
|
import "api/address.proto"; |
||||||
|
import "api/base.proto"; |
||||||
|
import "api/health_check.proto"; |
||||||
|
|
||||||
|
import "google/protobuf/wrappers.proto"; |
||||||
|
|
||||||
|
service EndpointDiscoveryService { |
||||||
|
// Translation of REST API to gRPC |
||||||
|
rpc Get(stream EndpointDiscoveryRequest) |
||||||
|
returns (stream EndpointDiscoveryResponse) { |
||||||
|
} |
||||||
|
|
||||||
|
// Advanced API to allow for multi-dimensional load balancing by remote |
||||||
|
// server. The steps are: |
||||||
|
// 1, The management server is configured with per cluster/zone/load metric |
||||||
|
// capacity configuration. The capacity configuration definition is |
||||||
|
// outside of the scope of this document. |
||||||
|
// 2. Envoy issues LoadBalanceRequest for a list of clusters it wants to load |
||||||
|
// balance to (instead of using a basic EndpointDiscoveryRequest). |
||||||
|
// 3. Once a connection establishes, the management server publishes |
||||||
|
// LoadBalanceResponse for all clusters requested in the first |
||||||
|
// LoadBalanceRequest. This message contains per cluster/per Locality |
||||||
|
// information. |
||||||
|
// 4. For each cluster, Envoy load balances incoming traffic to upstream hosts |
||||||
|
// based on per-zone weights and/or per-instance weights (if specified) |
||||||
|
// based on intra-zone LbPolicy. |
||||||
|
// 5. When upstream hosts reply, they optionally add header <define header |
||||||
|
// name> with ASCII representation of EndpointLoadMetricStats. |
||||||
|
// 6. Envoy aggregates load reports over the period of time given to it in |
||||||
|
// LoadBalanceResponse.load_reporting_interval. This includes aggregation |
||||||
|
// stats Envoy maintains by itself (total_requests, rpc_errors etc.) as |
||||||
|
// well as load metrics from upstream hosts. |
||||||
|
// 7. When the timer of load_reporting_interval expires, Envoy sends new |
||||||
|
// LoadBalanceRequest filled with load reports for each cluster. |
||||||
|
// 8. The management server uses the load reports from all reported Envoys |
||||||
|
// from around the world, computes global assignment and prepares traffic |
||||||
|
// assignment destined for each zone Envoys are located in. Goto 3. |
||||||
|
// TODO(htuch): Add @amb67's diagram. |
||||||
|
rpc BalanceLoad(stream LoadBalanceRequest) |
||||||
|
returns (stream LoadBalanceResponse) { |
||||||
|
} |
||||||
|
} |
||||||
|
|
||||||
|
message EndpointDiscoveryRequest { |
||||||
|
repeated string cluster_name = 1; |
||||||
|
Node node = 2; |
||||||
|
} |
||||||
|
|
||||||
|
message LbEndpoint { |
||||||
|
// Endpoint host address. |
||||||
|
Address address = 1; |
||||||
|
|
||||||
|
// Optional health status when known and supplied by EDS server. |
||||||
|
HealthStatus health_status = 2; |
||||||
|
|
||||||
|
// The optional canary status of the upstream host. Envoy uses the canary |
||||||
|
// status for various statistics and load balancing tasks documented |
||||||
|
// elsewhere. |
||||||
|
google.protobuf.BoolValue canary = 3; |
||||||
|
// The optional load balancing weight of the upstream host, in the range 1 - |
||||||
|
// 100. Envoy uses the load balancing weight in some of the built in load |
||||||
|
// balancers. All load balancing weights in a locality must sum to 100%. |
||||||
|
// The effective load balancing weight is load_balancing_weight multiplied by |
||||||
|
// the locality wide load_balancing_weight. If unspecified, each host is |
||||||
|
// presumed to have equal weight in a locality. |
||||||
|
// TODO(htuch): [V2-API-DIFF] Do we want all weights to sum to 100%, or to |
||||||
|
// instead use the sum of all weights in the denominator when computing |
||||||
|
// effective the weight ratio as done in v1? |
||||||
|
google.protobuf.UInt32Value load_balancing_weight = 4; |
||||||
|
} |
||||||
|
|
||||||
|
// All endpoints belonging to a Locality. |
||||||
|
message LocalityEndpoints { |
||||||
|
Locality locality = 1; |
||||||
|
repeated LbEndpoint lb_endpoints = 2; |
||||||
|
|
||||||
|
// Optional: Per region/zone/sub_zone weight - range 1-100. All load balancing |
||||||
|
// weights in a cluster must sum to 100%. If unspecified, each locality is |
||||||
|
// presumed to have equal weight in a cluster. |
||||||
|
// TODO(htuch): [V2-API-DIFF] Do we want all weights to sum to 100%, or to |
||||||
|
// instead use the sum of all weights in the denominator when computing |
||||||
|
// effective the weight ratio as done in v1? |
||||||
|
google.protobuf.UInt32Value load_balancing_weight = 3; |
||||||
|
} |
||||||
|
|
||||||
|
message EndpointDiscoveryResponse { |
||||||
|
repeated ClusterLoadBalance cluster_endpoints = 1; |
||||||
|
} |
||||||
|
|
||||||
|
// Example load report from a single request: |
||||||
|
// |
||||||
|
// [metric name, metric value] |
||||||
|
// * cpu_seconds, 0.7 |
||||||
|
// * flash_utilization, 75 |
||||||
|
// |
||||||
|
// When aggregating Envoy needs to count how many request's load reports |
||||||
|
// included each metric type, so Envoy can account for requests that don't |
||||||
|
// include that metric type. e.g.: |
||||||
|
// |
||||||
|
// [name, count, sum of values] |
||||||
|
// * cpu_seconds, 10, 17.5 |
||||||
|
// * flash_utilization, 5, 375 |
||||||
|
message EndpointLoadMetricStats { |
||||||
|
// Name of the metric; may be empty. |
||||||
|
string metric_name = 1; |
||||||
|
// Number of calls that finished and included this metric. |
||||||
|
uint64 num_requests_finished_with_metric = 2; |
||||||
|
// Sum of metric values across all calls that finished with this metric for |
||||||
|
// load_reporting_interval. |
||||||
|
double total_metric_value = 3; |
||||||
|
} |
||||||
|
|
||||||
|
// These are stats Envoy reports to GLB every so often. Report frequency is |
||||||
|
// defined by LoadBalanceResponse.interval |
||||||
|
// Stats per upstream region/zone and optionally per subzone |
||||||
|
message UpstreamLocalityStats { |
||||||
|
// Name of zone, region and optionally endpoint group this metrics was |
||||||
|
// collected from. Zone and region names could be empty if unknown. |
||||||
|
string Locality = 1; |
||||||
|
|
||||||
|
// The total number of requests sent by this Envoy since the last report. |
||||||
|
uint64 total_requests = 2; |
||||||
|
// The total number of unfinished requests |
||||||
|
uint64 total_requests_in_progress = 3; |
||||||
|
// The number of errors since the last report. |
||||||
|
enum TcpErrorType { |
||||||
|
TIMEOUT = 0; |
||||||
|
// TODO(htuch): Fill in additional TcpErrorType values. |
||||||
|
} |
||||||
|
// TODO(htuch): Ideally we would have the tcp_errors key be TcpErrorType, but |
||||||
|
// enums are not supported as map key types. Maybe make this a repeated |
||||||
|
// message with TcpErrorType x uint64 pairs. |
||||||
|
map<uint32, uint64> tcp_errors = 4; |
||||||
|
// HTTP status code, count |
||||||
|
map<uint32, uint64> http_errors = 5; |
||||||
|
// GRCP status code, count |
||||||
|
map<uint32, uint64> grpc_errors = 6; |
||||||
|
// The number of dropped requests since the last report. |
||||||
|
uint64 dropped_requests = 7; |
||||||
|
|
||||||
|
// Stats for multi-dimensional load balancing. |
||||||
|
repeated EndpointLoadMetricStats load_metric_stats = 8; |
||||||
|
} |
||||||
|
|
||||||
|
// Per cluster stats |
||||||
|
message ClusterStats { |
||||||
|
string cluster_name = 1; |
||||||
|
// Need at least one. |
||||||
|
repeated UpstreamLocalityStats upstream_locality_stats = 2; |
||||||
|
} |
||||||
|
|
||||||
|
message LoadBalanceRequest { |
||||||
|
Node node = 1; // zone/region where this Envoy runs |
||||||
|
repeated ClusterStats cluster_stats = 3; |
||||||
|
} |
||||||
|
|
||||||
|
// Each route from RDS will map to a single cluster or traffic split across |
||||||
|
// clusters using weights expressed in the RDS WeightedCluster. |
||||||
|
// |
||||||
|
// With EDS, each cluster is treated independently from a LB perspective, with |
||||||
|
// LB taking place between the Localities within a cluster and at a finer |
||||||
|
// granularity between the hosts within a locality. For a given cluster, the |
||||||
|
// effective weight of a host is its load_balancing_weight multiplied by the |
||||||
|
// load_balancing_weight of its Locality. |
||||||
|
message ClusterLoadBalance { |
||||||
|
string cluster_name = 1; |
||||||
|
repeated LocalityEndpoints endpoints = 2; |
||||||
|
// In the case where all endpoints for a particular zone/subzone are |
||||||
|
// unavailable/unhealthy, additional endpoints are given out for use in case |
||||||
|
// of catastrophic failure. They also have weights. |
||||||
|
repeated LocalityEndpoints failover_endpoints = 3; |
||||||
|
message Policy { |
||||||
|
// Percentage of traffic (0-100) that should be dropped. This |
||||||
|
// action allows protection of upstream hosts should they unable to |
||||||
|
// recover from an outage or should they be unable to autoscale and hence |
||||||
|
// overall incoming traffic volume need to be trimmed to protect them. |
||||||
|
// [V2-API-DIFF] This is known as maintenance mode in v1. |
||||||
|
double drop_overload = 1; |
||||||
|
} |
||||||
|
Policy policy = 4; |
||||||
|
} |
||||||
|
|
||||||
|
message LoadBalanceResponse { |
||||||
|
repeated ClusterLoadBalance clusters = 1; |
||||||
|
// The default is 10 seconds. |
||||||
|
Duration load_reporting_interval = 2; |
||||||
|
} |
Loading…
Reference in new issue