-
Notifications
You must be signed in to change notification settings - Fork 2k
nvidia-docker in centos #667
Comments
Can you provide the information in the issue template? https://github.com/NVIDIA/nvidia-docker/blob/master/.github/ISSUE_TEMPLATE.md#3-information-to-attach-optional-if-deemed-irrelevant At least the log and the output of |
==============NVSMI LOG============== Timestamp : Wed Mar 14 17:15:52 2018 Attached GPUs : 1 ======= uname -a ======= docker version Server: my system container have /etc/nvidia-container-runtime/config.toml debug line uncomment, no logs seems to be present in both container roofs and host.. |
for nvidia-container-cli -V I processed on a tmp test container, The system container has centos, Nvidia-driver is on the host, but should be accessible in the system container |
@olivier-dj did you solve the problem? Sorry for not answering earlier. |
Yes right excuse me, so to find a work-around I did some work on the Dockerfile of the system container, I did first mount / to /host in the container. Well it may seem strange in the sense that when we want to containerize an application, it can be to avoid side effects on the system. According to the member of the Atomic team, it's not that ugly because the system containers intend to provide sometimes critical services such as kernel modules, and in some case we would like to have access to the host system as if the module we provide was installed on the host system.
It was 2 months ago I don't remember exactly what was the bottleneck, but if I remember well if one of the links was missing it wouldn't work, (quite unexpected for the 32bit libraries). After that I probably had a problem with selinux and then after solving that It worked. |
hello, i'm using a fedora atomic host with a system container running docker-ce-17.03.02. I installed as well cuda and nvidia-docker2 in the container, and provided the system container via
atomic install --system --system-package no --storage ostree --name docker docker.io/olivenwk/centos-docker:17.03.2-cuda
(https://github.com/olivier-dj/atomic-system-containers/tree/update-docker-ce/docker-centos) the exports for cuda are supposed to be functionnal inside the system container. docker info says :Server Version: 17.03.2-ce
Runtimes: nvidia runc
Default Runtime: nvidia
Doing
docker run -it nginx bash
"works", it doesn't crash but if I do`docker run nvidia/cuda nvidia-smi'
I get :
/usr/bin/docker-current: Error response from daemon: oci runtime error: container_linux.go:247: starting container process caused "process_linux.go:337: running prestart hook 1 caused "error running hook: exit status 1, stdout: , stderr: exec command: [/usr/bin/nvidia-container-cli --load-kmods configure --device=all --compute --utility --require=cuda>=9.0 --pid=2377 /var/lib/docker/overlay2/cad8fca7bc7e5e606943ac111fea0ce4393e60ec7c01d8301cc1df737fd5a98c/merged]\nnvidia-container-cli: initialization error: driver error: failed to process request\n""
I'm quite sure /usr/bin/nvidia-container-cli is accessible inside the system container otherwise I would have "nvidia-container-cli\\\": executable file not found". The error message is not really clear :/
The text was updated successfully, but these errors were encountered: