Skip to content

Latest commit

Β 

History

History
72 lines (50 loc) Β· 3.8 KB

File metadata and controls

72 lines (50 loc) Β· 3.8 KB
title Qwen-Image ComfyUI Native Workflow Example
description Qwen-Image is a 20B parameter MMDiT (Multimodal Diffusion Transformer) model open-sourced under the Apache 2.0 license.
sidebarTitle Qwen-Image

import UpdateReminder from '/snippets/tutorials/update-reminder.mdx'

Qwen-Image is the first image generation foundation model released by Alibaba's Qwen team. It's a 20B parameter MMDiT (Multimodal Diffusion Transformer) model open-sourced under the Apache 2.0 license. The model has made significant advances in complex text rendering and precise image editing, achieving high-fidelity output for multiple languages including English and Chinese.

Model Highlights:

  • Excellent Multilingual Text Rendering: Supports high-precision text generation in multiple languages including English, Chinese, Korean, Japanese, maintaining font details and layout consistency
  • Diverse Artistic Styles: From photorealistic scenes to impressionist paintings, from anime aesthetics to minimalist design, fluidly adapting to various creative prompts

Related Links:

Qwen-Image Native Workflow Example

The models used in this document can be obtained from Huggingface or Modelscope

1. Workflow File

After updating ComfyUI, you can find the workflow file in the templates, or drag the workflow below into ComfyUI to load it. Qwen-image Text-to-Image Workflow

<a className="prose" target='_blank' href="https://raw.githubusercontent.com/Comfy-Org/workflow_templates/refs/heads/main/templates/image_qwen_image.json" style={{ display: 'inline-block', backgroundColor: '#0078D6', color: '#ffffff', padding: '10px 20px', borderRadius: '8px', borderColor: "transparent", textDecoration: 'none', fontWeight: 'bold'}}> <p className="prose" style={{ margin: 0, fontSize: "0.8rem" }}>Download JSON Workflow

2. Model Download

You can find all the models on Huggingface or Modelscope

Diffusion Model

Text Encoder

VAE

Model Storage Location

πŸ“‚ ComfyUI/
β”œβ”€β”€ πŸ“‚ models/
β”‚   β”œβ”€β”€ πŸ“‚ diffusion_models/
β”‚   β”‚   └── qwen_image_fp8_e4m3fn.safetensors
β”‚   β”œβ”€β”€ πŸ“‚ vae/
β”‚   β”‚   └── qwen_image_vae.safetensors
β”‚   └── πŸ“‚ text_encoders/
β”‚       └── qwen_2.5_vl_7b_fp8_scaled.safetensors

3. Complete the Workflow Step by Step

Step Guide

  1. Load qwen_image_fp8_e4m3fn.safetensors in the Load Diffusion Model node
  2. Load qwen_2.5_vl_7b_fp8_scaled.safetensors in the Load CLIP node
  3. Load qwen_image_vae.safetensors in the Load VAE node
  4. Set image dimensions in the EmptySD3LatentImage node
  5. Enter your prompts in the CLIP Text Encoder (supports English, Chinese, Korean, Japanese, Italian, etc.)
  6. Click Queue or press Ctrl+Enter to run