You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
What happened:
perhaps this is a coincidence, but after updating from ami version amazon-eks-node-1.19-v20210322 to version
amazon-eks-node-1.19-v20210628(ami-0c2ca9cd067f101bc) in region eu-west-3, after some time, an error with container amazon-k8s-cni:v1.7.5-eksbuild.1 appeared on one of the 9 nodes:
Type Reason Age From Message
---- ------ ---- ---- -------
Normal Scheduled 3m4s default-scheduler Successfully assigned kube-system/aws-node-n9dmj to ip-10-10-10-10.eu-west-3.compute.internal
Normal Pulled 3m3s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Successfully pulled image "602401143452.dkr.ecr.eu-west-3.amazonaws.com/amazon-k8s-cni:v1.7.5-eksbuild.1" in 127.941902ms
Normal Pulling 3m3s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Pulling image "602401143452.dkr.ecr.eu-west-3.amazonaws.com/amazon-k8s-cni-init:v1.7.5-eksbuild.1"
Normal Pulled 3m3s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Successfully pulled image "602401143452.dkr.ecr.eu-west-3.amazonaws.com/amazon-k8s-cni-init:v1.7.5-eksbuild.1" in 146.91813ms
Normal Created 3m3s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Created container aws-vpc-cni-init
Normal Started 3m3s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Started container aws-vpc-cni-init
Normal Started 3m2s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Started container aws-node
Warning Unhealthy 2m56s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:20:59.792Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 2m46s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:09.775Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 2m36s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:19.771Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 2m26s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:29.762Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 2m16s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:39.770Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 2m6s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:49.762Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 116s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:21:59.769Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 112s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Liveness probe failed: {"level":"info","ts":"2021-07-02T09:22:03.775Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Warning Unhealthy 106s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Readiness probe failed: {"level":"info","ts":"2021-07-02T09:22:09.766Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Normal Killing 92s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Container aws-node failed liveness probe, will be restarted
Warning Unhealthy 86s (x4 over 102s) kubelet, ip-10-10-10-10.eu-west-3.compute.internal (combined from similar events): Readiness probe failed: {"level":"info","ts":"2021-07-02T09:22:29.784Z","caller":"/usr/local/go/src/runtime/proc.go:203","msg":"timeout: failed to connect service \":50051\" within 1s"}
Normal Pulling 82s (x2 over 3m3s) kubelet, ip-10-10-10-10.eu-west-3.compute.internal Pulling image "602401143452.dkr.ecr.eu-west-3.amazonaws.com/amazon-k8s-cni:v1.7.5-eksbuild.1"
Normal Created 81s (x2 over 3m3s) kubelet, ip-10-10-10-10.eu-west-3.compute.internal Created container aws-node
Normal Pulled 81s kubelet, ip-10-10-10-10.eu-west-3.compute.internal Successfully pulled image "602401143452.dkr.ecr.eu-west-3.amazonaws.com/amazon-k8s-cni:v1.7.5-eksbuild.1" in 504.321983ms
logs:
{"level":"info","ts":"2021-07-02T09:14:11.287Z","caller":"entrypoint.sh","msg":"Install CNI binary.."}
{"level":"info","ts":"2021-07-02T09:14:11.305Z","caller":"entrypoint.sh","msg":"Starting IPAM daemon in the background ... "}
{"level":"info","ts":"2021-07-02T09:14:11.307Z","caller":"entrypoint.sh","msg":"Checking for IPAM connectivity ... "}
restarting and recreating the container didn't help, the problem was solved only by deleting the node.
How to reproduce it (as minimally and precisely as possible):
update the ami to the version amazon-eks-node-1.19-v20210628(ami-0c2ca9cd067f101bc)
Anything else we need to know?:
previously, there were no such errors, so I decided to create a ticket
What happened:
perhaps this is a coincidence, but after updating from ami version amazon-eks-node-1.19-v20210322 to version
amazon-eks-node-1.19-v20210628(ami-0c2ca9cd067f101bc) in region eu-west-3, after some time, an error with container amazon-k8s-cni:v1.7.5-eksbuild.1 appeared on one of the 9 nodes:
logs:
restarting and recreating the container didn't help, the problem was solved only by deleting the node.
How to reproduce it (as minimally and precisely as possible):
update the ami to the version amazon-eks-node-1.19-v20210628(ami-0c2ca9cd067f101bc)
Anything else we need to know?:
previously, there were no such errors, so I decided to create a ticket
Environment:
aws eks describe-cluster --name <name> --query cluster.platformVersion
): eks.4aws eks describe-cluster --name <name> --query cluster.version
): 1.19uname -a
): Linux ip-10-10-10-10.eu-west-3.compute.internal 5.4.117-58.216.amzn2.x86_64 Template is missing source_ami_id in the variables section #1 SMP Tue May 11 20:50:07 UTC 2021 x86_64 x86_64 x86_64 GNU/Linuxcat /etc/eks/release
on a node):The text was updated successfully, but these errors were encountered: