What is the correct way of using Forge for Flux checkpoints (CUDA OOM error) #1844
nitinmukesh
started this conversation in
General
Replies: 1 comment
-
And I am surprised to learn that changing GPU weight to 13000 worked. Though it defaulted to the max VRAM automatically. Funny thing is it used only 5.x GB VRAM and generate the image. Last time it crashed and CUDA OOM error while using 7.8 GB. Can someone explain plz |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
So I was using Auto1111 all the time before switching to Forge. I never get CUDA OOM error before while using Auto1111.
Something which I learned (from variety of tools like Auto1111, Diffsynth studio, Talking head, etc) is that there is GPU VRAM and then Shared memory from system RAM. First the GPU VRAM memory is utilized fully and then the rest will be spilled to shared RAM and if both are full and more is needed than you get CUDA OOM error.
However in Forge, it is not fully utilizing the VRAM and starts utilizing shared RAM. To support this check this image

4.8 Gb of VRAM is used and then started to use shared. So I got CUDA OOM error on 7.5 GB / 15.9 GB, which is not correct. It should be for e.g. 14 GB / 15.9 GB and application request for another 2 GB one should get CUDA OOM error.
Please find the console log and settings I am using. What am I doing wrong, Please advice.
I think GPU weight plays a role in this but looking at the value it seems correct, use 7xxx VRAM. Or maybe my understanding is incorrect.
Beta Was this translation helpful? Give feedback.
All reactions