-
Notifications
You must be signed in to change notification settings - Fork 385
lifecycle-sidecar at 100% of CPU limit #515
Comments
Hey Kyle, are you seeing this on your connect pods or on the mesh gateway?
We can definitely make this configurable but because it should be the same
for everyone we might instead just want to bump it up. Our testing showed
it had lots of room but maybe that does depend on workload.
Can you show us your graphs or kubectl top output as well please?
…On Thu, Jun 25, 2020, 7:28 PM Kyle Purdon ***@***.***> wrote:
I'm seeing right at peak memory usage for the lifecycle sidecar. Would it
be possible to increase, or allow configuration of the resource requests
for this.
I *think*
https://github.com/hashicorp/consul-helm/blob/e2fb73ca3fa00074ceec26f6b4b15dd860a094ad/templates/mesh-gateway-deployment.yaml#L359-L365
is the spot that would need to allow for configuration.
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<#515>, or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AAH4RPI6CZCZA34GID7E6S3RYQBUJANCNFSM4OI5MFFA>
.
|
This is on the connect enabled pods, no mesh gateway. Here is a chart showing the CPU limit utilization for all lifecycle containers for the last hour: Here is a single top for one of the pods:
Happy to provide any additional metrics that may be useful. |
Okay thank you for this information. We're going to look to bump this up and we're also looking at another issue related to resource settings and OOM. If you need a workaround right now you'd need to bump these up yourself: and build a custom consul-k8s image. To be clear this is a high priority for us and we're working on a fix as we speak. |
Will be addressed by #533 |
This bugfix is available in 0.23.0. |
Awesome @lkysow ... quick question. Is the same multi-step upgrade process required for using a newer helm chart version if the underlying consul version has not changed? |
Hey Kyle, it depends on whether the consul client daemonset pods will end up being restarted by the helm upgrade. The release (you should actually use 0.23.1, there was a TLS bug we just patched) changes the default version of consul-k8s to 0.17.0 in order to get this bugfix. That shouldn't affect the client daemonset unless you have ACLs enabled. If you do, then consul-k8s is actually used as an init container in the client daemonset and so bumping the Docker image version will trigger a client daemonset restart and so for a no-downtime upgrade you would need to follow the multi-step upgrade process. There's no built-in way to helm to see what will be updated but there is a helm diff plugin: https://github.com/databus23/helm-diff
If the output contains a line about the daemonset like
Then the client will be updated. |
I'm seeing right at 100% of the requested limit CPU usage for the lifecycle sidecar. Would it be possible to increase, or allow configuration of the resource requests for this.
I think
consul-helm/templates/mesh-gateway-deployment.yaml
Lines 359 to 365 in e2fb73c
The text was updated successfully, but these errors were encountered: