Skip to content

Commit

Permalink
upstream: Add ability to disable host selection during panic
Browse files Browse the repository at this point in the history
Previously, when in a panic state, requests would be routed to all
hosts. In some cases it is instead preferable to not route any requests.
Add a configuration option for zone-aware load balancers which switches
from routing to all hosts to no hosts.

Closes envoyproxy#7550.

Signed-off-by: James Forcier <jforcier@grubhub.com>
  • Loading branch information
James Forcier committed Aug 22, 2019
1 parent f12adac commit b73e501
Show file tree
Hide file tree
Showing 6 changed files with 96 additions and 10 deletions.
4 changes: 4 additions & 0 deletions api/envoy/api/v2/cds.proto
Original file line number Diff line number Diff line change
Expand Up @@ -539,6 +539,10 @@ message Cluster {
// * :ref:`runtime values <config_cluster_manager_cluster_runtime_zone_routing>`.
// * :ref:`Zone aware routing support <arch_overview_load_balancing_zone_aware_routing>`.
google.protobuf.UInt64Value min_cluster_size = 2;

// If set to true, Envoy will not consider any hosts when the cluster is in panic mode.
// Instead, the cluster will fail all requests as if all hosts are unhealthy.
bool disable_cluster_on_panic = 3;
}
// Configuration for :ref:`locality weighted load balancing
// <arch_overview_load_balancing_locality_weighted_lb>`
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,8 @@ Panic threshold

During load balancing, Envoy will generally only consider available (healthy or degraded) hosts in
an upstream cluster. However, if the percentage of available hosts in the cluster becomes too low,
Envoy will disregard health status and balance amongst all hosts. This is known as the *panic
threshold*. The default panic threshold is 50%. This is
Envoy will disregard health status and balance either amongst all hosts or no hosts. This is known
as the *panic threshold*. The default panic threshold is 50%. This is
:ref:`configurable <config_cluster_manager_cluster_runtime>` via runtime as well as in the
:ref:`cluster configuration <envoy_api_field_Cluster.CommonLbConfig.healthy_panic_threshold>`.
The panic threshold is used to avoid a situation in which host failures cascade throughout the
Expand All @@ -20,8 +20,12 @@ disregards panic thresholds and continues to distribute traffic load across prio
the algorithm described :ref:`here <arch_overview_load_balancing_priority_levels>`.
However, when normalized total availability drops below 100%, Envoy assumes that there are not enough
available hosts across all priority levels. It continues to distribute traffic load across priorities,
but if a given priority level's availability is below the panic threshold, traffic will go to all hosts
in that priority level regardless of their availability.
but if a given priority level's availability is below the panic threshold, traffic will go to all
(or no) hosts in that priority level regardless of their availability.

There are two modes Envoy can choose from when in a panic state: traffic will either be sent to all
hosts, or will be sent to no hosts (and therefore will always fail). This is configured in the
:ref:`cluster configuration <envoy_api_field_Cluster.CommonLbConfig.ZoneAwareLbConfig.disable_cluster_on_panic>`.

The following examples explain the relationship between normalized total availability and panic threshold.
It is assumed that the default value of 50% is used for the panic threshold.
Expand Down
1 change: 1 addition & 0 deletions docs/root/intro/version_history.rst
Original file line number Diff line number Diff line change
Expand Up @@ -42,6 +42,7 @@ Version history
certificate validation context.
* upstream: added network filter chains to upstream connections, see :ref:`filters<envoy_api_field_Cluster.filters>`.
* upstream: use p2c to select hosts for least-requests load balancers if all host weights are the same, even in cases where weights are not equal to 1.
* upstream: add :ref:`disable_cluster_on_panic <envoy_api_field_Cluster.CommonLbConfig.ZoneAwareLbConfig.disable_cluster_on_panic>` to allow failing all requests to a cluster during panic state
* zookeeper: parse responses and emit latency stats.

1.11.1 (August 13, 2019)
Expand Down
23 changes: 18 additions & 5 deletions source/common/upstream/load_balancer_impl.cc
Original file line number Diff line number Diff line change
Expand Up @@ -282,7 +282,8 @@ ZoneAwareLoadBalancerBase::ZoneAwareLoadBalancerBase(
routing_enabled_(PROTOBUF_PERCENT_TO_ROUNDED_INTEGER_OR_DEFAULT(
common_config.zone_aware_lb_config(), routing_enabled, 100, 100)),
min_cluster_size_(PROTOBUF_GET_WRAPPED_OR_DEFAULT(common_config.zone_aware_lb_config(),
min_cluster_size, 6U)) {
min_cluster_size, 6U)),
disable_cluster_on_panic_(common_config.zone_aware_lb_config().disable_cluster_on_panic()) {
ASSERT(!priority_set.hostSetsPerPriority().empty());
resizePerPriorityState();
priority_set_.addPriorityUpdateCb(
Expand Down Expand Up @@ -552,7 +553,11 @@ ZoneAwareLoadBalancerBase::hostSourceToUse(LoadBalancerContext* context) {
// If the selected host set has insufficient healthy hosts, return all hosts.
if (per_priority_panic_[hosts_source.priority_]) {
stats_.lb_healthy_panic_.inc();
hosts_source.source_type_ = HostsSource::SourceType::AllHosts;
if (disable_cluster_on_panic_) {
hosts_source.source_type_ = HostsSource::SourceType::NoHosts;
} else {
hosts_source.source_type_ = HostsSource::SourceType::AllHosts;
}
return hosts_source;
}

Expand Down Expand Up @@ -586,9 +591,13 @@ ZoneAwareLoadBalancerBase::hostSourceToUse(LoadBalancerContext* context) {

if (isGlobalPanic(localHostSet())) {
stats_.lb_local_cluster_not_ok_.inc();
// If the local Envoy instances are in global panic, do not do locality
// based routing.
hosts_source.source_type_ = sourceType(host_availability);
// If the local Envoy instances are in global panic, and we should not disable the cluster, do
// not do locality based routing.
if (disable_cluster_on_panic_) {
hosts_source.source_type_ = HostsSource::SourceType::NoHosts;
} else {
hosts_source.source_type_ = sourceType(host_availability);
}
return hosts_source;
}

Expand All @@ -610,6 +619,8 @@ const HostVector& ZoneAwareLoadBalancerBase::hostSourceToHosts(HostsSource hosts
return host_set.healthyHostsPerLocality().get()[hosts_source.locality_index_];
case HostsSource::SourceType::LocalityDegradedHosts:
return host_set.degradedHostsPerLocality().get()[hosts_source.locality_index_];
case HostsSource::SourceType::NoHosts:
return dummy_empty_host_vector;
default:
NOT_REACHED_GCOVR_EXCL_LINE;
}
Expand Down Expand Up @@ -696,6 +707,8 @@ void EdfLoadBalancerBase::refresh(uint32_t priority) {
HostsSource(priority, HostsSource::SourceType::LocalityDegradedHosts, locality_index),
host_set->degradedHostsPerLocality().get()[locality_index]);
}
add_hosts_source(HostsSource(priority, HostsSource::SourceType::NoHosts),
dummy_empty_host_vector);
}

HostConstSharedPtr EdfLoadBalancerBase::chooseHostOnce(LoadBalancerContext* context) {
Expand Down
7 changes: 6 additions & 1 deletion source/common/upstream/load_balancer_impl.h
Original file line number Diff line number Diff line change
Expand Up @@ -180,14 +180,16 @@ class ZoneAwareLoadBalancerBase : public LoadBalancerBase {
LocalityHealthyHosts,
// Degraded hosts for locality @ locality_index.
LocalityDegradedHosts,
// No hosts in the host set.
NoHosts,
};

HostsSource() = default;

HostsSource(uint32_t priority, SourceType source_type)
: priority_(priority), source_type_(source_type) {
ASSERT(source_type == SourceType::AllHosts || source_type == SourceType::HealthyHosts ||
source_type == SourceType::DegradedHosts);
source_type == SourceType::DegradedHosts || source_type == SourceType::NoHosts);
}

HostsSource(uint32_t priority, SourceType source_type, uint32_t locality_index)
Expand Down Expand Up @@ -231,6 +233,8 @@ class ZoneAwareLoadBalancerBase : public LoadBalancerBase {
*/
const HostVector& hostSourceToHosts(HostsSource hosts_source);

const HostVector dummy_empty_host_vector;

private:
enum class LocalityRoutingState {
// Locality based routing is off.
Expand Down Expand Up @@ -300,6 +304,7 @@ class ZoneAwareLoadBalancerBase : public LoadBalancerBase {

const uint32_t routing_enabled_;
const uint64_t min_cluster_size_;
const bool disable_cluster_on_panic_;

struct PerPriorityState {
// The percent of requests which can be routed to the local locality.
Expand Down
59 changes: 59 additions & 0 deletions test/common/upstream/load_balancer_impl_test.cc
Original file line number Diff line number Diff line change
Expand Up @@ -539,6 +539,39 @@ TEST_P(FailoverTest, PriorityUpdatesWithLocalHostSet) {
EXPECT_EQ(tertiary_host_set_.hosts_[0], lb_->chooseHost(nullptr));
}

// Test that extending the priority set with an existing LB causes the correct updates when the
// cluster is configured to disable on panic.
TEST_P(FailoverTest, PriorityUpdatesWithLocalHostSetDisableOnPanic) {
host_set_.hosts_ = {makeTestHost(info_, "tcp://127.0.0.1:80")};
failover_host_set_.hosts_ = {makeTestHost(info_, "tcp://127.0.0.1:81")};
common_config_.mutable_zone_aware_lb_config()->set_disable_cluster_on_panic(true);

init(false);
// With both the primary and failover hosts unhealthy, we should select no host.
EXPECT_EQ(nullptr, lb_->chooseHost(nullptr));

// Update the priority set with a new priority level P=2 and ensure the host
// is chosen
MockHostSet& tertiary_host_set_ = *priority_set_.getMockHostSet(2);
HostVectorSharedPtr hosts(new HostVector({makeTestHost(info_, "tcp://127.0.0.1:82")}));
tertiary_host_set_.hosts_ = *hosts;
tertiary_host_set_.healthy_hosts_ = tertiary_host_set_.hosts_;
HostVector add_hosts;
add_hosts.push_back(tertiary_host_set_.hosts_[0]);
tertiary_host_set_.runCallbacks(add_hosts, {});
EXPECT_EQ(tertiary_host_set_.hosts_[0], lb_->chooseHost(nullptr));

// Now add a healthy host in P=0 and make sure it is immediately selected.
host_set_.healthy_hosts_ = host_set_.hosts_;
host_set_.runCallbacks(add_hosts, {});
EXPECT_EQ(host_set_.hosts_[0], lb_->chooseHost(nullptr));

// Remove the healthy host and ensure we fail back over to tertiary_host_set_
host_set_.healthy_hosts_ = {};
host_set_.runCallbacks({}, {});
EXPECT_EQ(tertiary_host_set_.hosts_[0], lb_->chooseHost(nullptr));
}

// Test extending the priority set.
TEST_P(FailoverTest, ExtendPrioritiesUpdatingPrioritySet) {
host_set_.hosts_ = {makeTestHost(info_, "tcp://127.0.0.1:80")};
Expand Down Expand Up @@ -829,6 +862,32 @@ TEST_P(RoundRobinLoadBalancerTest, MaxUnhealthyPanic) {
EXPECT_EQ(3UL, stats_.lb_healthy_panic_.value());
}

// Test that no hosts are selected when disable_cluster_on_panic is enabled.
TEST_P(RoundRobinLoadBalancerTest, MaxUnhealthyPanicDisableOnPanic) {
hostSet().healthy_hosts_ = {makeTestHost(info_, "tcp://127.0.0.1:80"),
makeTestHost(info_, "tcp://127.0.0.1:81")};
hostSet().hosts_ = {
makeTestHost(info_, "tcp://127.0.0.1:80"), makeTestHost(info_, "tcp://127.0.0.1:81"),
makeTestHost(info_, "tcp://127.0.0.1:82"), makeTestHost(info_, "tcp://127.0.0.1:83"),
makeTestHost(info_, "tcp://127.0.0.1:84"), makeTestHost(info_, "tcp://127.0.0.1:85")};

common_config_.mutable_zone_aware_lb_config()->set_disable_cluster_on_panic(true);

init(false);
EXPECT_EQ(nullptr, lb_->chooseHost(nullptr));

// Take the threshold back above the panic threshold.
hostSet().healthy_hosts_ = {
makeTestHost(info_, "tcp://127.0.0.1:80"), makeTestHost(info_, "tcp://127.0.0.1:81"),
makeTestHost(info_, "tcp://127.0.0.1:82"), makeTestHost(info_, "tcp://127.0.0.1:83")};
hostSet().runCallbacks({}, {});

EXPECT_EQ(hostSet().healthy_hosts_[0], lb_->chooseHost(nullptr));
EXPECT_EQ(hostSet().healthy_hosts_[1], lb_->chooseHost(nullptr));

EXPECT_EQ(1UL, stats_.lb_healthy_panic_.value());
}

// Ensure if the panic threshold is 0%, panic mode is disabled.
TEST_P(RoundRobinLoadBalancerTest, DisablePanicMode) {
hostSet().healthy_hosts_ = {};
Expand Down

0 comments on commit b73e501

Please sign in to comment.