Replies: 3 comments 1 reply
-
I will take a look next week or so. You can repost again later if I forget this |
Beta Was this translation helpful? Give feedback.
1 reply
-
Bump for repost. |
Beta Was this translation helpful? Give feedback.
0 replies
-
This was posted on reddit today I hope it gets implemented, it could increase speed in models like Flux and be faster |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I've looked at using TensorRT in the past, but there are a number of issues that has me hesitant atm. While looking for updates I found this interesting and thought I'd share.
Recently stumbled across Diffusers + Torchao
"We provide end-to-end inference and experimental training recipes to use torchao with diffusers in this repo. We demonstrate 53.88% speedup on Flux.1-Dev* and 21% speedup on CogVideoX-5b when comparing compiled quantized models against their standard bf16 counterparts**."
Beta Was this translation helpful? Give feedback.
All reactions