WAN 2.2 converting to GGUF #788
Replies: 4 comments 6 replies
-
|
Beta Was this translation helpful? Give feedback.
-
Thanks for pointing me to the wan branch. I tried switching to that branch and build it using cmake per instruction (clone recursive then switch to wan branch). But it error out with the following messages: D:\tools\stable-diffusion.cpp\model.cpp(1061,17): error C2027: use of undefined type 'gguf_tensor_shape' [D:\tools\stab D:\tools\stable-diffusion.cpp\model.cpp(1072,17): error C3861: 'gguf_init_from_file_ext': identifier not found [D:\tool D:\tools\stable-diffusion.cpp\ggml_extend.hpp(883,18): error C3861: 'ggml_conv_3d': identifier not found [D:\tools\stab D:\tools\stable-diffusion.cpp\wan.hpp(76,17): error C3861: 'ggml_pad_ext': identifier not found [D:\tools\stable-diffus D:\tools\stable-diffusion.cpp\ggml_extend.hpp(883,18): error C3861: 'ggml_conv_3d': identifier not found [D:\tools\stab Not sure if I did something wrong during the cloning process. |
Beta Was this translation helpful? Give feedback.
-
After updating the submodule, it started to convert but it always stop / crash at the same spot of a WAN 2.2 model. I ran with the -v flag (assuming it's verbose?) but no explicit error messages pop up. Any suggestions? |
Beta Was this translation helpful? Give feedback.
-
For the lightning lora to work we are limited by the supported quants seen here: I have updated the docs to point out that in the lora.md That leaves us with just Q8_0 supported when using the Wan2.2 GGUF found at: It would be nice to have support for a bit less heavy quants like Q5_K_S that I have been able to generate images with on my Amd 7800xt 16GB in ComfyUI |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I saw the author's pull request over at Llama CPP which incorporated handling of WAN. While the pull request was merged, the quantize binary (exe file) does not recognize the wan format. So I clone the latest sd.cpp but the convert function also failed with <model.cpp:1177 - invalid tensor 'model.diffusion_model.patch_embedding.weight'>.
So my questions are:
Note, I already tried the ComfyUI GGUF approach and that also failed.
Beta Was this translation helpful? Give feedback.
All reactions