Replies: 29 comments 55 replies
-
Currently, the speed of Nvidia version 531.79 is normal during testing. I'm not sure if the old version 531.68.61 will be faster. I'll test it later. |
Beta Was this translation helpful? Give feedback.
-
Tested with versions 531.79-68-64-41-29, the speed difference does not exceed 10 seconds. The 532 version has a speed reduction of 20-50%, and the 535 version has a speed reduction of over 9000%. This situation may be encountered on graphics cards with less than 12GB of memory. |
Beta Was this translation helpful? Give feedback.
-
in my test, [email protected] work well |
Beta Was this translation helpful? Give feedback.
-
My generations werent really slower but I had a lot of artifacting when using tiled vae or img2img. Its gone now, I rolled back by installing an older studio driver |
Beta Was this translation helpful? Give feedback.
-
Driver ver 535 is a shame, A1111 runs tooooo slowwwwww on that driver version. I had to get back to 532. Problem solved. |
Beta Was this translation helpful? Give feedback.
-
It would be a great feature if efficient shared VRAM usage could be achieved by the driver alone. The premise is that Nvidia provides a switch and at the same time cooperates with the program to optimize the use of logic. |
Beta Was this translation helpful? Give feedback.
-
Nvidia's goal/solution is likely that we adapt our workflow to use their olive tensor implementation (and don't run projects large enough to dip into system memory). |
Beta Was this translation helpful? Give feedback.
-
In fact, a FP32 model generates with 2x the IT/s than FP16 that i pruned from it. So the driver indeed screwed with speeds. An sadly, the newest drivers has diablo 4 optimizations, so i'd have to have SD on a separate windows install lol.
I would rather it dip into system memory than using pagefile on my poor nvme whenever video memory runs out from a generation heh. |
Beta Was this translation helpful? Give feedback.
-
DON'T UPGRADE TO v535.** |
Beta Was this translation helpful? Give feedback.
-
Hello, Sorry if this not in the correct place. I had installed Konya_ss with the new installer. Very good. In middle I install the NVIDIA library : that run very fast my 3090. Can be of interest to incorporate to A1111 Hope this info be good for this community. THANKS FOR ALL ! |
Beta Was this translation helpful? Give feedback.
-
Could also be the source of sudden NaN errors when using Adetailer, maybe triggered by mixing Adetailer's models with ControlNet or the mixing of FP16 with FP32. It's a shame because I'd rather it dip into shared memory when going above VRAM limit at the temporary cost of speed, instead of just stopping altogether, but without the extra bugs. |
Beta Was this translation helpful? Give feedback.
-
This doesn't matter much because Windows 11 insider drivers are mostly limited to insider builds and their changes are all over the place with a mix of older than stable branch code and new, despite having higher build numbers than regular drivers, but Nvidia Insider driver 545.29 doesn't seem to have this memory management change. |
Beta Was this translation helpful? Give feedback.
-
I've been noticing Stable Diffusion rendering slowdowns since updating to the latest nvidia GRD but it gets more complicated than that. I had heard from a reddit post that rolling back to 531.79 would solve the speed reduction and it did but a reboot undid that and returned me to slow-land. This was especially noticeable as I had finally gotten xformers (0.0.17) to load without errors and instead of a speed boost I was getting painfully unusably slow render speeds. Turns out reinstalling the drivers without reboot, regardless whether the latest GRD or the rollback version people have been suggesting, fixes the problem temporarily, at least for the current desktop session, and then the problems return after reboot. The workaround for this is to reinstall nvidia drivers prior to working with stable diffusion, but we shouldn't have to do this. The silver lining is that the latest nvidia drivers do indeed include the memory management improvements that eliminate OOM errors by hitting shared gpu (system) RAM instead of crashing out with OOM, but at the expense of speed when it's fallen back to system memory. This general slowdown does not appear related to the slowdowns one would expect when falling back to slower memory, but the slowdowns are similar. You can watch this play out on the performance tab of Task Manager when memory overflow is the reason for your slowdowns. I still have no idea why failing to reinstall nvidia drivers prior to an SD session results in speeds comparable to using system RAM, or if they are even related. On my system I get over 15it/s with a blank prompt at 512 square using default A1111 settings, euler a, 20 steps, CFG 7. This is only after a reinstall of the latest drivers from nvidia, which were already installed. This is repeatable. Is it unique to my system or are others noticing the same things with their A1111 install? When it's running without first reinstalling nvidia drivers (without reboot), I get more like 1.2it/s, which as you might guess is really noticeable. One more thing. I was told by someone that the new version of torch makes xformers kind of irrelevant but it doesn't seem that way from my perspective. I tested some renders without it turned on in launch args and they seem a little slower than expected but it seems like xformers is still boosting things when I have it turned on. I have to test whether a fresh reinstall of nvidia drivers changes that assessment, but I know from testing that it has accelerated things back to 15it/s with --xformers on (and loaded). I will update this when I've had a chance to test that scenario against xformers being turned off and see if the performance returns the same as it does when reinstalling nvidia drivers with it turned off. UPDATE: I just removed --xformers from my a1111 startup args and did a test render post-nvidia-reinstall and got about 10-11it/s, so it seems xformers (I'm running 0.0.17) is not obsoleted by the torch upgrade, at least not entirely. I still get 15-16it/s with it turned on, as long as the drivers have been reinstalled during the current session, otherwise things slow down almost entirely, to 1.2it/s or the even worse flipside of that, seconds per iteration. I put it back on and 15+ it/s as expected. I don't know if that's fast or slow but at least it's back to the speeds I got used to before I ever got xformers to load properly. Until reboot at least. UPDATE: Well, now I'm getting what I would consider normal render speeds without having to reinstall the nvidia driver so this is where I throw up my hands and shrug. Device name DESKTOP-QH9LO7U |
Beta Was this translation helpful? Give feedback.
-
Nvidia is aware of performance issues with the new memory management and will provide a fix in future driver updates. |
Beta Was this translation helpful? Give feedback.
-
What driver does run well? I had way better performance im march than now. I went back to 531 version, but it did not help. |
Beta Was this translation helpful? Give feedback.
-
RTX 30 series here Installed 536.40 and immediately noticed that image generation was taking 4 or even more times longer than the previous driver. But the interesting part, was that I also noticed that the GPU simply wasn't being stressed nearly as much either. I have the RGB bar on my GPU set to change colors based on Temperature, and normally my GPU temp sky rockets during image generation. And with the new driver the temps stayed in the low range for the GPU without ever increasing. Rolled back to 536.23 and the performance gain was immediate along with the temps that I was used to seeing. Edit: Only noticed after my post that this was in reference to the driver released even further back. Will roll back even further to see if I get more of a boost than rolling back to 536.23. |
Beta Was this translation helpful? Give feedback.
-
Dunno what's going on exactly. I have the latest Nvidia drivers and Comfy absolutely blasts SDXL renders out. Faster than 1.5 on A1111. |
Beta Was this translation helpful? Give feedback.
-
Some have suggested that the shared memory usage appears at lower required
memory usages than 531. This hasn’t been addressed in drivers yet but is on
their radar.
Depending on how things evolve with SDXL we might just hit a point when low
memory cards aren’t viable anymore.
…On Thu, Jul 20, 2023 at 14:08 gunnerman1 ***@***.***> wrote:
I did some of my own tests with 536.* and 531.79. Performance differences
were trivial in the short tests I did.
Where I think people may be seeing the slowdown is on my rig, using
531.79, SD will not use shared gpu memory and will simply throw an OOM
error when dedicated gou memory is exhausted. 536 drivers, shared memory
will be used if needed. This allows for much larger gens etc however, once
SD bleeds into shared memory, even by a little bit, generation time will
suffer severely. People should check their task manager and ensure the jobs
are not hitting shared gpu memory if you are on newer drivers before
concluding the driver is a problem.
Mid gen screenshots to show. First is before HiRes fix, second is after,
shared memory is being used and will slow things down. The large spike is
the VAE tiling.
[image: image]
<https://user-images.githubusercontent.com/11324979/254983968-629a9be5-378e-4d3d-bc82-df0d700d869e.png>
[image: image]
<https://user-images.githubusercontent.com/11324979/254984491-68a1fd74-1707-40bf-bb7e-68d531c53ef9.png>
—
Reply to this email directly, view it on GitHub
<#11062 (reply in thread)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AOKLG4Z65DFTLTZ73OKRB5LXRF62NANCNFSM6AAAAAAY4S3UA4>
.
You are receiving this because you commented.Message ID:
<AUTOMATIC1111/stable-diffusion-webui/repo-discussions/11062/comments/6502774
@github.com>
|
Beta Was this translation helpful? Give feedback.
-
RTX 3060 12GB user here. As soon as i removed I know many people say you should never use |
Beta Was this translation helpful? Give feedback.
-
Driver 536.99 acknowledges the issue and says a future update should fix it. "This driver implements a fix for creative application stability issues seen during heavy memory usage. We’ve observed some situations where this fix has resulted in performance degradation when running Stable Diffusion and DaVinci Resolve. This will be addressed in an upcoming driver release. [4172676]" |
Beta Was this translation helpful? Give feedback.
-
lol what's a shame Nvidia! |
Beta Was this translation helpful? Give feedback.
-
Excerpt from the freshly released 537.13 driver: [DaVinci Resolve] This driver implements a fix for creative application stability issues So it seems they tried to implement a fix for other apps like SD? My GPU just died, but can someone test and see if things improved? Thanks! |
Beta Was this translation helpful? Give feedback.
-
Still living with 531.79 till this day, It seems NVIDIA wont even bother since this is on the lower end of GPUs 😑 |
Beta Was this translation helpful? Give feedback.
-
Does anyone test the latest 537.42 (WHQL) ? Is the issue fixed? |
Beta Was this translation helpful? Give feedback.
-
I've installed studio 537.58 and have generated 5 images at 680x512 with hires fix at 2.2 with various samplers and a bunch of loras and plug ins (SAG, Anti-burn, Tiled Vae) and all of them have been consistently the same speed and without any noticeable slowdown or difference from 531 so far. GTX4050TI laptop with 4gb of VRAM and 64GB RAM. Will keep on this driver and see how it goes. Would appreciate other opinions, I'm now paranoid and twitchy and utterly beholden to subejctiveness where this issue is concerned :-) |
Beta Was this translation helpful? Give feedback.
-
UpdateUsers are reporting the issue have been fixed in 537.58 |
Beta Was this translation helpful? Give feedback.
-
After testing, 537.58 did not fix this issue. |
Beta Was this translation helpful? Give feedback.
-
I tested 545.84 (game ready) which resolved the problem for me. It was not fixed for 537.58 |
Beta Was this translation helpful? Give feedback.
-
seem NVIDIA update it, workaround for 546.01 please check https://nvidia.custhelp.com/app/answers/detail/a_id/5490 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Update (2023-10-31)
This issue should now be entirely resolved. NVIDIA has made a help article to disable the system memory fallback behavior. Please upgrade to the latest driver (546.01) and follow the guide on their website: https://nvidia.custhelp.com/app/answers/detail/a_id/5490
Update (2023-10-17)
there seems to be some reports saying that the issue is still not fixed
comments
#11063 (comment)
#11063 (comment)
Update (2023-10-14)
Users are reporting the issue have been fixed in 537.58
#11063 (comment)
some users have reported some issues related to the latest (535) Nvidia drivers
nVidia drivers change in memory management vladmandic#1285
#11050 (comment)
if you have been experiencing generation slowdowns or getting stuck, consider downgrading to driver version 531 or below
NVIDIA Driver Downloads
Beta Was this translation helpful? Give feedback.
All reactions