tutorials/video/wan/wan2_2 #291
Replies: 38 comments 48 replies
-
14B takes too long on a 3090. (5 minutes for 1 step) 5B gives an error on the VAE. Any solutions? |
Beta Was this translation helpful? Give feedback.
-
Using the default 5B TI2V workflow I'm getting some artifacts, particulary often with I2V (T2V is sometimes producing them, but are less noticeable). This artifacts appear like flickering giant pixels all over the generated video. I've only changed the aspect ratio to portrait and lowered framecount so I get quicker gens. Anyone knows if this may cause known trouble or if this issue is common across some configs? Thank you in advance. |
Beta Was this translation helpful? Give feedback.
-
On the 14B workflows I'm getting an error of: Requested to load WAN21 Is this just me running out of VRAM? I have 16 GB Also noticed on the image to video 14B workflow the Json is loading the FP8 models but the text says to download the FP16 models. I'm a noob so maybe I'm just misunderstanding what this is telling us. |
Beta Was this translation helpful? Give feedback.
-
This guide says to use the 2.1 VAE for 14B model... Is that a mistake? |
Beta Was this translation helpful? Give feedback.
-
When i tried the 14b i2v workflow using the fp8_e4m3fn_fast i get this error: Exception during fp8 op: CUDA error: CUBLAS_STATUS_NOT_SUPPORTED when calling Switching to fp8_e5m2 made it run however i'm not sure if this will make the results worse, I'm running it on a RTX 4090 and one step is taking roughly 2.3minutes |
Beta Was this translation helpful? Give feedback.
-
is there any reason that the secon samples is on fixed value? |
Beta Was this translation helpful? Give feedback.
-
Ich erhalte beim Workflow folgende Fehlermeldung: KSamplerAdvanced |
Beta Was this translation helpful? Give feedback.
-
The workflows are coming out for me as just an image, not a video. Not sure what's going on. |
Beta Was this translation helpful? Give feedback.
-
definetly needs lora speedup |
Beta Was this translation helpful? Give feedback.
-
Error: Given groups=1, weight of size [5120, 36, 1, 2, 2], expected input[1, 32, 31, 160, 90] to have 36 channels, but got 32 channels instead |
Beta Was this translation helpful? Give feedback.
-
5b model i2v 960x1280 (!) resolution takes just 3:07 with 20 steps. really fast. it is base fp16. so fp8 should be less then 2min and fp4 just around 1min or even less. with lower resolution ofc faster. i tried lower resolution it takes less then 2min. but quality much worst |
Beta Was this translation helpful? Give feedback.
-
the negative prompt in the 5b workflow is in chinese, does it matter? running it through deepl gave me "Vibrant colors, overexposed, static, blurry details, subtitles, style, artwork, painting, scene, static, overall grayish, poor quality, low quality, JPEG compression artifacts, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, deformed limbs, fused fingers, static scene, cluttered background, three legs, many people in the background, walking backward" is that what they're supposed to be? or should i leave them in chinese? what about if i want to add more, does the language matter? |
Beta Was this translation helpful? Give feedback.
-
Very cool but on the "Wan2.2 14B T2V Text-to-Video Workflow Example" i got a stall with a 4090, 64gb ram, 9950x3d .. Shortening the length gave me an error supposedly caused by EmptyHunyuanLatentVideo. I lowered the resolution in the hopes it would work but keep getting AI told me that there should be a "EmptyWAN2LatentVideo " |
Beta Was this translation helpful? Give feedback.
-
Can I use wan2.2 with same Causvid Lora ? |
Beta Was this translation helpful? Give feedback.
-
Can I run with multiple GPUs? |
Beta Was this translation helpful? Give feedback.
-
3090 with 64 ram (no sageattention) works well and pretty fast creating 480x480 @ ComfyUI: v0.3.47 using I2V -14Bfp8.scaled if use WAN21_T2V_14B_lightx2V_cfg_step_distill_lora. Only 6 steps needed. |
Beta Was this translation helpful? Give feedback.
-
currently training a Lora... and its training lightening fast! kinda crazy! very cool. for 1000 epoch's its only going to take about 15 minutes! |
Beta Was this translation helpful? Give feedback.
-
anyone have loras that just stay? even after removing the whole lora loader node? |
Beta Was this translation helpful? Give feedback.
-
Im confused why the workflow tutorial is using the 14B models for Image to Video when the 5B models claims to be able to do both Image to Video and Text to Video |
Beta Was this translation helpful? Give feedback.
-
does it work with rocm (linux) on amd rx 9070xt? Just wondering if and how anyone made it work, I Will test too tonight. |
Beta Was this translation helpful? Give feedback.
-
Amazing results with an very old Hp600 (12 cores ... from2007) and RTX3060 12go. |
Beta Was this translation helpful? Give feedback.
-
getting error message "UnboundLocalError: cannot access local variable 'clip_vision_output' where it is not associated with a value" any ideas on what I'm doing wrong? am using the exact workflow |
Beta Was this translation helpful? Give feedback.
-
I have this error with FL2V : cannot access local variable 'clip_vision_output' where it is not associated with a value |
Beta Was this translation helpful? Give feedback.
-
For those with low VRAM (8GB)... Start with ComfyUI Official Guide WAN 2.2 Ti2V: Relevant file locations: "C:\ComfyUI\models\diffusion_models\wan2.2_ti2v_5B_fp16.safetensors" Continue with this tutorial (Minute >=27) works faster: Use Quanstack quantized GGUF models for WAN 2.2: Replace Load Difussion Model node, with Unet Loader (GGUF) to load the GGUF model. Relevant file locations: |
Beta Was this translation helpful? Give feedback.
-
Anyone come across this issue? 4090 + 64gb ram. I ran updates across all nodes and have the latest NVIDIA studio drivers |
Beta Was this translation helpful? Give feedback.
-
wan2.2_i2v_high_noise_14B_fp16 Have this error |
Beta Was this translation helpful? Give feedback.
-
Hey I am using the Wan 2.2 workflow with 14b fp8 scaled high and low noise with the 4steps lora lightx2v lora. The speed is impressive but I seem to get brown static and weird visual abnormalities especially when I turn up the quality. I was wondering why this may be happening? |
Beta Was this translation helpful? Give feedback.
-
First, thank you so much for the awesome guide and example workflow! 🙏 From the Wan2.2 14B FLF2V Workflow, I had a quick question about the fp8_scaled + 4 steps LoRA part. On the second Ksampler, I see start_at_step = 2 and end_at_step = 10000. Should the end_at_step actually be set to 4 instead, since the other one with 4 steps LoRA uses end_at_step = 4? |
Beta Was this translation helpful? Give feedback.
-
<3 Any idea why? Thank you! |
Beta Was this translation helpful? Give feedback.
-
I found that the generated video will be forced to have a same image to be the first and last frame of the video if I use a >81 value for length in EmptyHunyuanLatentVideo node, and the first frames visibly have higher contrast and saturation. I can't find out that if this is an expected behavior or not. Has anybody also encountered such behavior? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
tutorials/video/wan/wan2_2
Official User Guide for WAN 2.2 Video Generation workflow example in ComfyUI
https://docs.comfy.org/tutorials/video/wan/wan2_2
Beta Was this translation helpful? Give feedback.
All reactions