How to get started adapting LightX2V techniques to Text to Image Models #212
Unanswered
MS00-GitIt
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi team—thanks for LightX2V, it’s great work. I’ve read the docs and README and see official support for HunyuanVideo, Wan 2.1/2.2, SkyReels-V2-DF, and CogVideoX1.5.
My goal: explore whether LightX2V methods (e.g., step distillation, quantization, attention kernels, offloading) can be adapted to image models like Skyworks UniPic (MIT), SDXL, or Chroma (Flux Schnell variant). I realize these are image T2I models (not video), so I’m asking:
My environment (for context): Linux, CUDA 12.x, Python 3.10; GPU: RTX 3090 24GB VRAM.
Happy to prototype and contribute a PR or notes if you can point me to the right extension points.
Links I consulted:
Beta Was this translation helpful? Give feedback.
All reactions