- 
                Notifications
    You must be signed in to change notification settings 
- Fork 6.5k
Closed
Labels
bugSomething isn't workingSomething isn't working
Description
Describe the bug
When I use bfloat16 or float16 VAE computing type - I have error of different input and model types.
On encode step.
Reproduction
# --- VAE / VAE ---
    vae = AutoencoderKLWan.from_pretrained(
        "Wan-AI/Wan2.1-VACE-14B-diffusers",
        subfolder="vae",
        torch_dtype=torch.float16,
    )
    vae.enable_tiling(TARGET_HEIGHT, int(TARGET_WIDTH // 1.5))
    # --- Text Encoder / Кодировщик текста ---
    # [TODO - 0001] No support at 10.08.2025 - https://github.com/huggingface/transformers/issues/40067
    # text_encoder = UMT5EncoderModel.from_pretrained(
    #     "city96/umt5-xxl-encoder-gguf",
    #     gguf_file="umt5-xxl-encoder-Q8_0.gguf",
    #     torch_dtype=torch.float16,
    # )
    # --- Transformer / Трансформер ---
    transformer = WanVACETransformer3DModel.from_single_file(
        "https://huggingface.co/QuantStack/Wan2.1_T2V_14B_FusionX_VACE-GGUF/blob/main/Wan2.1_T2V_14B_FusionX_VACE-Q6_K.gguf",
        quantization_config=GGUFQuantizationConfig(
            compute_dtype=torch.float16
        ),
        torch_dtype=torch.float16,
    )
    # --- Pipeline assembly / Сборка пайплайна ---
    pipe = WanVACEPipeline.from_pretrained(
        "Wan-AI/Wan2.1-VACE-14B-diffusers",
        vae=vae,
        #text_encoder=text_encoder, # [TODO - 0001]
        transformer=transformer,
        torch_dtype=torch.float16
    )
    # --- Scheduler / Планировщик ---
    pipe.scheduler = UniPCMultistepScheduler.from_config(pipe.scheduler.config, flow_shift=flow_shift)
    # Sage
    pipe.transformer.set_attention_backend("sage")  # stable for masks / стабильно для масок
    log.info("Attention backend set to 'sage' / Бэкенд внимания установлен: 'sage'")
    # VAE memory save
    pipe.enable_model_cpu_offload()
    log.info("WAN pipeline is ready / Пайплайн WAN готов")
    return pipe
Logs
System Info
- 🤗 Diffusers version: 0.35.0.dev0
- Platform: Windows-11-10.0.26100-SP0
- Running on Google Colab?: No
- Python version: 3.12.10
- PyTorch version (GPU?): 2.7.1+cu128 (True)
- Flax version (CPU?/GPU?/TPU?): not installed (NA)
- Jax version: not installed
- JaxLib version: not installed
- Huggingface_hub version: 0.34.3
- Transformers version: 4.54.1
- Accelerate version: 1.1.0
- PEFT version: 0.15.2
- Bitsandbytes version: 0.46.0
- Safetensors version: 0.5.3
- xFormers version: 0.0.31.post1
- Accelerator: NVIDIA GeForce RTX 4090, 24564 MiB
- Using GPU in script?:
- Using distributed or parallel set-up in script?: No
- Using: accelerate launch
Who can help?
J4BEZ
Metadata
Metadata
Assignees
Labels
bugSomething isn't workingSomething isn't working