Skip to content

ZImage output is just noise after hack to make zimage work in fp16 #11087

@mamei16

Description

@mamei16

Custom Node Testing

Expected Behavior

The output should somewhat resemble what it was before when I ran the model and "manual cast: FP32" was printed in the logs.

Actual Behavior

The output is just noise.

Image

Steps to Reproduce

I can reproduce it by simply running the following workflow:

Image

Debug Logs

Checkpoint files will always be loaded safely.
Total VRAM 12272 MB, total RAM 31494 MB
pytorch version: 2.4.1+rocm6.1
AMD arch: gfx1030
ROCm version: (6, 1)
Set vram state to: NORMAL_VRAM
Device: cuda:0 AMD Radeon RX 6800M : native
Enabled pinned memory 29919.0
Using sub quadratic optimization for attention, if you have memory or speed issues try using: --use-split-cross-attention
Python version: 3.11.13 (main, Jun  5 2025, 13:12:00) [GCC 11.2.0]
ComfyUI version: 0.3.76
ComfyUI frontend version: 1.33.10
[Prompt Server] web root: /home/user/envs/stable_diffusion/lib/python3.11/site-packages/comfyui_frontend_package/static
Total VRAM 12272 MB, total RAM 31494 MB
pytorch version: 2.4.1+rocm6.1
AMD arch: gfx1030
ROCm version: (6, 1)
Set vram state to: NORMAL_VRAM
Device: cuda:0 AMD Radeon RX 6800M : native
Enabled pinned memory 29919.0
Skipping loading of custom nodes
Context impl SQLiteImpl.
Will assume non-transactional DDL.
No target revision found.
Starting server

To see the GUI go to: http://127.0.0.1:8188
got prompt
Using split attention in VAE
Using split attention in VAE
VAE load device: cuda:0, offload device: cpu, dtype: torch.float32
CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cpu, dtype: torch.float16
Requested to load ZImageTEModel_
loaded completely; 11028.80 MB usable, 7672.25 MB loaded, full load: True
model weight dtype torch.float16, manual cast: None
model_type FLOW
unet missing: ['norm_final.weight']
Requested to load Lumina2
loaded partially; 4795.20 MB usable, 4720.20 MB loaded, 7019.35 MB offloaded, 75.00 MB buffer reserved, lowvram patches: 0
100%|██████████████████████████████████████████████████████████| 8/8 [01:21<00:00, 10.14s/it]
Requested to load AutoencodingEngine
0 models unloaded.
loaded completely; 4773.80 MB usable, 319.75 MB loaded, full load: True
Prompt executed in 114.16 seconds

Other

I am fairly certain that the issue stems from commit #11057.

Metadata

Metadata

Assignees

No one assigned

    Labels

    Potential BugUser is reporting a bug. This should be tested.

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions