You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Added instructions for converting HuggingFace checkpoints to Megatron format and vice versa, including necessary commands and notes on exported checkpoints.
**Note**: Current inference path is single-GPU. Parallel inference is not yet supported.
149
149
150
+
151
+
---
152
+
153
+
### 🔄 Checkpoint Converting (optional)
154
+
155
+
If you plan to fine-tune Wan using a pre-trained model, you must first convert the HuggingFace checkpoint (e.g., `Wan-AI/Wan2.1-T2V-1.3B-Diffusers`) into the Megatron format. The provided script supports bidirectional conversion, allowing you to move between HuggingFace and Megatron formats as needed.
**Note**: The exported checkpoint from Megatron to HuggingFace (`/workspace/checkpoints/hf_checkpoints/wan_1_3b_hf`) contains only the DiT transformer weights. To run inference, you still require the other pipeline components (VAE, text encoders, etc.).
178
+
To assemble a functional inference directory:
179
+
- Duplicate the original HF checkpoint directory.
180
+
- Replace the `./transformer` folder in that directory with your newly exported `/transformer` folder.
0 commit comments