Skip to content

Conversation

@stefanasandei
Copy link

Hello, I was trying to finetune Chroma1-HD on my setup, RTX 3090 and 32gb RAM. However, I was getting stuck during model loading, because the RAM was getting filled, resulting in my computer freezing every time.

In the ChromaModel class, the safetensors file is loaded first into the CPU, to be quantized. The Chroma checkpoint is too big to be loaded in full precision within 32gb RAM, so I changed the global dtype to be bfloat16 during the initial loading, and then the normal dtypes afterwards.

With this configuration, I can easily finetune Chroma with my setup.

I also fixed a bug related to FakeTextEncoder during loading (related issue: #405).

@stefanasandei stefanasandei changed the title Enable Chroma loading within 32gb RAM & 24gb VRAM Enable Chroma finetune within 32gb RAM & 24gb VRAM Oct 29, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant