Skip to content

Conversation

@Koratahiu
Copy link
Contributor

@Koratahiu Koratahiu commented Nov 14, 2025

This PR implements the timestep sampling method from:
A Closer Look at Time Steps is Worthy of Triple Speed-Up for Diffusion Model Training.

Claims 3x faster pretraining at same quality:

image

Usage

  • Set timestep distribution to SPEED

⚠️ Notes

  • Validated for pretraining only; finetuning impact unknown, but the core concept may apply.

TODO

  • To be tested
  • Minimal change: The current approach modifies _get_timestep_discrete and requires betas/sigmas, which is not ideal.

I don't like this approach, but until we fine a better minimal way
@miasik
Copy link

miasik commented Nov 14, 2025

Should I expect better quality for the same steps amount for fine-tuning?
Or, what should I pay attention to in order to test it?

@O-J1
Copy link
Collaborator

O-J1 commented Nov 14, 2025

@Koratahiu The image you included is 404.

@miasik
Copy link

miasik commented Nov 14, 2025

I usually use "debiased estimation" as loss weight function. Should I set it to constant for using SpeeD?

@Koratahiu
Copy link
Contributor Author

Koratahiu commented Nov 14, 2025

Should I expect better quality for the same steps amount for fine-tuning? Or, what should I pay attention to in order to test it?

Yeah, if it works, then it should converge faster in the same number of steps.

I usually use "debiased estimation" as loss weight function. Should I set it to constant for using SpeeD?

The paper mentions that it’s compatible with loss weight functions (e.g., p2, min-SNR, debiased estimation, etc.), and in their official repo, they set the loss weight function to p2 by default.

@Koratahiu
Copy link
Contributor Author

Closed, as #1225 seems to have a better and theoretically-sound approach (and it achieves that with minimal code).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants