[QUESTION/HELP] Installing NVIDIA GPU operator on a k3d cluster #1458
Replies: 2 comments 1 reply
-
Hey! |
Beta Was this translation helpful? Give feedback.
-
Hello @iwilltry42 , thank you for your reply. I modified the helm chart of gpu operator to disable the symllink creator and was able to avoid the above issue. But now I am running into a new error Error: failed to generate container "7ef3f1c73ce130dc46badbdef38af202bd8e52e495b0e285f823dfd320f4288e" spec: failed to generate spec: path "/run/nvidia/driver" is mounted on "/run" but it is not a shared or slave mount Can you tell me if this is an error with my underlying docker infrastructure? Specifically I am not sure how to mount the /run folder as a shared mount |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hello,
I am trying to install NVIDIA GPU operator on a K3d cluster. I have a GPU cluster setup according to the docs(https://k3d.io/v5.6.3/usage/advanced/cuda/) and am able to access the GPU using any pods created in the cluster.
But when I try to install NVIDIA GPU operator as described in the install guide (https://docs.nvidia.com/datacenter/cloud-native/gpu-operator/latest/getting-started.html#) the driver validator is stuck in a error loop. Here are the logs from the pod
Can you help me solve this issue?
Beta Was this translation helpful? Give feedback.
All reactions