-
Notifications
You must be signed in to change notification settings - Fork 545
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Removing hostNetwork: true breaks upgrades #1127
Comments
Tracking in #1111. |
Closing the issue as it is tracked in 1111. |
@mskanth972: Closing this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
/reopen Hola @fraenkel !!! |
@dims: Reopened this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
Hi @fraenkel, thanks for pointing that out. I don't think the different common name should have any affect here, although I will double check. Are you seeing an error like the following when you describe the efs-csi-node Pod which performed the mount?
Can you follow our trouble shooting guide to enable debug mode and pull driver + efs-utils logs? Thanks for the help debugging, |
No, the issue is that you see no error other than the mount point hangs if you attempt to use it. |
So I enabled debug and the only thing you see from the efs-plugin logs are:
the efs_utils log:
this pattern repeats over and over for each mount. |
if I delete a pod that has a mounted volume, the only change is a single line in the log
and the pod is stuck in Terminating. If I go into the efs plugin and attempt to umount, it just hangs. A force, says the target is busy. |
dmesg has a ton of
|
kubelet logs show
this repeats until I force deleted the mount. |
We've also ran into this issue, we found that restarting/recreating the nodes fixed the problem, but an in place upgrade of a minor version should definitely not be causing EFS mounts to no longer work. I've found it happens when upgrading the helm chart from 2.4.2->2.4.9 and also 2.4.7->2.4.9 |
We ran into the same problem. My theory is that the kernel establishes the NFS mount to
Even upgrading to v2.5.0 of the Helm chart (which reenabled |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
/remove-lifecycle rotten |
The Kubernetes project currently lacks enough contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle stale |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues. This bot triages un-triaged issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /lifecycle rotten |
The Kubernetes project currently lacks enough active contributors to adequately respond to all issues and PRs. This bot triages issues according to the following rules:
You can:
Please send feedback to sig-contributor-experience at kubernetes/community. /close not-planned |
@k8s-triage-robot: Closing this issue, marking it as "Not Planned". In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes-sigs/prow repository. |
/kind bug
What happened?
Upgrading from 1.5.8 to 1.6.0, all of our existing attached storage no longer work.
What you expected to happen?
Pods continue to have attached storage.
How to reproduce it (as minimally and precisely as possible)?
Deploy 1.5.8 or 1.5.9.
Deploy a pod that attaches EFS.
Upgrade to 1.6.0.
Anything else we need to know?:
If we rollback, things work again.
The one change we do notice is that the CN is different based on the hostNetwork = true/false. You either get the node's hostname or the pod's hostname depending.
Environment
kubectl version
):Please also attach debug logs to help us better diagnose
The only thing we notice is that the mount-watchdog.log will kill the stunnel every 5 minutes due to an unhealthy tunnel.
The text was updated successfully, but these errors were encountered: