Load Phi 3 small on Nvidia Tesla V100 - Flash Attention #1956
Replies: 1 comment 4 replies
-
Hey, this seems to be unrelated to PEFT, right? So please open a discussion for transformers instead, which looks like the right place.
ONNX is not for training, so it's not an option here. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi,
I would like to inquire about the possibility of uploading and fine tuning a Phi 3 8k small. When I load the model, I get an error about missing Flash attention. If I want to install the given package, I get this error :
But I have the required version of pytorch and CUDA (torch 2.3.1 and cuda 12.1)
Is it because I am using a Tesla V100 graphics card? Is there any way to load the model also with this graphics card?
I found this in the documentation for the Phi 3 mini on Huggingface:
Does this also apply to the Phi3 Small 8k?? Beacause when I try to load it, the error occurs
Or should I try the ONNX version or it is just for inference?
Thank you.
Beta Was this translation helpful? Give feedback.
All reactions