From 2e9b0f404c53ee221ae02ac142f910b9cfffea9b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Ond=C5=99ej=20Benkovsk=C3=BD?= Date: Mon, 23 Aug 2021 13:40:26 +0200 Subject: [PATCH] remove unused healthcheck metrics MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Signed-off-by: Ondřej Benkovský --- README.md | 3 --- metrics.go | 12 ------------ 2 files changed, 15 deletions(-) diff --git a/README.md b/README.md index 8b42fcc..0dd193c 100644 --- a/README.md +++ b/README.md @@ -36,9 +36,6 @@ If monitoring is enabled (via the *prometheus* plugin) then the following metric * `coredns_fanout_request_duration_seconds{to}` - duration per upstream interaction. * `coredns_fanout_request_count_total{to}` - query count per upstream. * `coredns_fanout_response_rcode_count_total{to, rcode}` - count of RCODEs per upstream. -* `coredns_fanout_healthcheck_failure_count_total{to}` - number of failed health checks per upstream. -* `coredns_fanout_healthcheck_broken_count_total{}` - counter of when all upstreams are unhealthy, - and we are randomly (this always uses the `random` policy) spraying to an upstream. Where `to` is one of the upstream servers (**TO** from the config), `rcode` is the returned RCODE from the upstream. diff --git a/metrics.go b/metrics.go index 900215c..69c5598 100644 --- a/metrics.go +++ b/metrics.go @@ -43,16 +43,4 @@ var ( Buckets: plugin.TimeBuckets, Help: "Histogram of the time each request took.", }, []string{"to"}) - HealthcheckFailureCount = promauto.NewCounterVec(prometheus.CounterOpts{ - Namespace: plugin.Namespace, - Subsystem: "fanout", - Name: "healthcheck_failure_count_total", - Help: "Counter of the number of failed healthchecks.", - }, []string{"to"}) - HealthcheckBrokenCount = promauto.NewCounter(prometheus.CounterOpts{ - Namespace: plugin.Namespace, - Subsystem: "fanout", - Name: "healthcheck_broken_count_total", - Help: "Counter of the number of complete failures of the healthchecks.", - }) )