Replies: 2 comments
-
This does not look like anything podman has to do with, I read that like you app requires a cuda capability your GPU simply does not have. As such I do not believe this is a podman bug and converting it to a discussion |
Beta Was this translation helpful? Give feedback.
-
https://github.com/Dao-AILab/flash-attention says
|
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Issue Description
Running the command
podman run -it --rm --security-opt=label=disable --device nvidia.com/gpu=all --shm-size 1g -p 8080:80 -v /opt/lorax/data:/data:Z ghcr.io/predibase/lorax:latest --model-id mistralai/Mistral-7B-Instruct-v0.1
I get the error
GPU with CUDA capability 7 0 is not supported for Flash Attention V2
I am running on an RHEL9 with podman version:
Steps to reproduce the issue
Steps to reproduce the issue
podman run -it --rm --security-opt=label=disable --device nvidia.com/gpu=all --shm-size 1g -p 8080:80 -v /opt/lorax/data:/data:Z ghcr.io/predibase/lorax:latest --model-id mistralai/Mistral-7B-Instruct-v0.1
GPU with CUDA capability 7 0 is not supported for Flash Attention V2
Describe the results you received
error
GPU with CUDA capability 7 0 is not supported for Flash Attention V2
Describe the results you expected
No startup error
podman info output
Podman in a container
No
Privileged Or Rootless
Rootless
Upstream Latest Release
No
Additional environment details
Additional environment details
Additional information
I ran what is described on https://docs.nvidia.com/datacenter/cloud-native/container-toolkit/latest/cdi-support.html:
and also, I updated /etc/subuid and /etc/subgid for running rootless
Full stack trace of the error:
nvidia-smi provides:
Beta Was this translation helpful? Give feedback.
All reactions