Proper way use MULTI GPUs with DEEPSPEED when each GPU does not have enough VRAM? #1369
Unanswered
atipasvanund
asked this question in
Q&A
Replies: 1 comment
-
You should be able to using zero3 with offload. Alternatively , make sure to adjust optim etc. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
What's the proper way to fine-tune on MULTI GPUs with DEEPSPEED when each GPU does not have enough VRAM? For example If I wish to use 4 x RTX 4090 to fine-tune a 7B model without LORA or Quantization. Is this possible? I tried zero1.json all the way to zero3_bf16.json and get various error, including out of CUDA memory. Can someone please help me?
Beta Was this translation helpful? Give feedback.
All reactions