Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Losing contact with an instance after a while #2504

Closed
thehappycoder opened this issue Mar 30, 2022 · 5 comments
Closed

Losing contact with an instance after a while #2504

thehappycoder opened this issue Mar 30, 2022 · 5 comments
Labels

Comments

@thehappycoder
Copy link

Describe the bug
I've installed Kubernetes on 3 instances on an Apple M1 machine, and after connecting these instances in a Kubernetes cluster via kubeadm join, I lost contact to all 3 instances.

I've managed to regain access by restarting multipass:

sudo launchctl unload /Library/LaunchDaemons/com.canonical.multipassd.plist && sudo launchctl load /Library/LaunchDaemons/com.canonical.multipassd.plist

But then the instances freeze after a while. It also happens when I run only 1 instance.

Logs
I get this in /Library/Logs/Multipass/multipassd.log when I lose contact with my instances:

[2022-03-31T07:28:21.508] [debug] [k8s-worker2] QMP: {"timestamp": {"seconds": 1648675701, "microseconds": 508459}, "event": "NIC_RX_FILTER_CHANGED", "data": {"path": "/machine/unattached/device[7]/virtio-backend"}}

Additional info
Apple M1
macOS 12.3
multipass 1.8.1+mac

@Saviq
Copy link
Collaborator

Saviq commented Mar 31, 2022

Hi @thehappycoder, we've not seen this before I don't think. This message suggests the kernel reinitializes the virtual network (you should see it when it's first booting as well). I was actually unable to get that message even unloading the network module, are you certain the message only shows when networking goes down?

If you launch an instance without k8s, does it happen as well?

@thehappycoder
Copy link
Author

thehappycoder commented Mar 31, 2022

@Saviq

Yes, it may have something to do with k8s running or some related commands I ran on an instance because a brand new instance seems to be not affected with the freeze happens. Here are the commands that I ran 1643991933135-Building a Kubernetes Cluster.pdf

Also I am not sure if it's normal or not, but memory usage of qemu processes seems suspiciously high for some instances
Screen Shot 2022-03-31 at 8 33 04 pm
.

@Saviq
Copy link
Collaborator

Saviq commented Mar 31, 2022

@thehappycoder there's a lot of networking foo happening there, it's quite possible something there is messing it up.

You could try using a bridged network for the k8s bits, I just posted a pre-release package here: #2364 (comment)

I'll close this, then.

@Saviq Saviq closed this as completed Mar 31, 2022
@Saviq Saviq added invalid and removed bug labels Mar 31, 2022
@Saviq
Copy link
Collaborator

Saviq commented Mar 31, 2022

Ah @thehappycoder on memory use, I recently explained some of it here:

#2494 (comment)

@thehappycoder
Copy link
Author

Turns out something in the software running on these virtual machines is messing up SSH because I also experience this when I connect to virtual machines via SSH that are running in UTM.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants