Replies: 9 comments 4 replies
-
The information you provided is not enough. At present, I can only tell you that the SDE sampler is slower, and Euler a is twice as fast as him. |
Beta Was this translation helpful? Give feedback.
-
I have the same type of card, 3060 with 12GB vram, 32Gb memory, i5-8400 Default setup from this repo on Manjaro Linux, no upgrading to Torch 2.0 or anything, just clean setup about 2 week ago. |
Beta Was this translation helpful? Give feedback.
-
This is what I'm getting on my system. I have a 3070 though, but maybe it can give some insight? RTX 3070 | Ryzen 7 3700x | ram 16Gb 8.5-9it/s - 512x512 px | 20 Steps | CFG 7 | Euler a |
Beta Was this translation helpful? Give feedback.
-
Hey I'm running into your exact problem and am lost, did you ever find a fix ? |
Beta Was this translation helpful? Give feedback.
-
is everyone using a batch size of 1? I have the same problem here 3060 12g, 32g RAM, i9-9900k, windows 11, I've turned off GPU acceleration in windows and Chrome. A fresh clone of master which includes pytorch 2.0.1, and nothing else added. I get 1.2it/s running 512 x 512, 20 steps, euler a, , no negative prompt and only "cat" as a prompt. I only get 5.8it/s with a batch size of 1, with a batch size of 8 i only get 1.2 it/s I even tried command line args below as suggested elsewhere with no change |
Beta Was this translation helpful? Give feedback.
-
The recent updates have been experiencing this issue, with a random decrease in speed of 5-10 times,Previously, a 960 resolution image took 30 seconds, but now it takes 5-10 minutes |
Beta Was this translation helpful? Give feedback.
-
This thread has brought to my attention that I've been getting low performance as well on my 3060 12GB. I kept these commandline arguments: --opt-sdp-attention --opt-channelslast |
Beta Was this translation helpful? Give feedback.
-
Back in April, I encountered the same issue as you did, but no matter how much I tried to make changes, nothing worked. So, I gave up and decided to use an older version instead. These past few days, I decided to give it another try and found that the same issues persisted. The problems I faced back then were twofold: first, I couldn't achieve a resolution of 1920x1280, and second, the performance was extremely slow. To tackle these problems, I used the following parameters: set COMMANDLINE_ARGS= --xformers --lowvram --medvram --precision full --no-half --no-half-vae --disable-nan-check --opt-split-attention-v1 --upcast-sampling --opt-channelslast --theme dark --autolaunch After testing, I confirmed that these settings were effective, but it resulted in extremely slow processing times, taking about 20 minutes. Therefore, I decided to revert to the parameters I have been using all along, and I ran the test again using the following parameters: set COMMANDLINE_ARGS= --medvram --xformers --force-enable-xformers --always-batch-cond-uncond --opt-channelslast --no-hashing --disable-nan-check --api --xformers-flash-attention --opt-split-attention --no-half-vae This time, the results were similar in terms of speed to what I experienced with the older version. Time taken: 1m 4.92sTorch active/reserved: 11238/11582 MiB, Sys VRAM: 12288/12288 MiB (100.0%) |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hey, if people are getting ~14it/s with RTX 3070, why is my 3060 juicing only 0.85it/s? It's over 15x slower and I still don't know why. What are some checks I could do and maybe possible fixes?
RTX 3060 12GB | Ryzen 5 5600X | 16gb 2666hz
pyTorch 2.0+cu118 | xFormers 0.0.18
512x512 px | 20 Steps | CFG 7 | DPM++ SDE Karras
Beta Was this translation helpful? Give feedback.
All reactions