-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GPU addon does not fully set up GPU access #272
Comments
Hello, OS: Fedora 39 Server My Set-Up Procedure
I hope this Helps. |
@ca-scribner in case the of the ref 2 is not related to the GPU add-on I am manually installing the toolkit, rather than having Microk8s install it Now, since I am also installing Kubeflow on top of this And since, I am manually installing the toolkit I am manually editing a file that I am not sure how persistent this change will be across microk8s updates sudo vi /var/snap/microk8s/current/args/containerd-template.toml
[..]
# default_runtime_name = "${RUNTIME}"
default_runtime_name = "nvidia"
sudo systemctl restart snap.microk8s.daemon-containerd.service |
Summary
A few Charmed Kubeflow users have reported that, after doing
microk8s enable gpu
, the GPUs are not exposed to their workloads unless they also setnvidia
as the default runtime in thecontainerd-template.toml
.I'm not sure if this is a new bug or something that has happened for a while, but it has been raised twice (1, 2) this month.
What Should Happen Instead?
For a machine that has an nvidia gpu, using
microk8s enable gpu
(or the newernvidia
addon, for recent microk8s versions) would fully set up the GPU for use by podsReproduction Steps
See this thread for more details (sorry, I'm creating this to report someone else's issue)
Introspection Report
Sorry, I'm creating this to report someone else's issue and don't have the report
Can you suggest a fix?
No
Are you interested in contributing with a fix?
No, but will cc others who might
The text was updated successfully, but these errors were encountered: