You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Fine-tuning Kontext on the T2I task can be useful when working with specific styles/subjects where it may not
295
295
perform as expected.
296
296
297
+
Image-guided fine-tuning (I2I) is also supported. To start, you must have a dataset containing triplets:
298
+
299
+
* Condition image
300
+
* Target image
301
+
* Instruction
302
+
303
+
[kontext-community/relighting](https://huggingface.co/datasets/kontext-community/relighting) is a good example of such a dataset. If you are using such a dataset, you can use the command below to launch training:
More generally, when performing I2I fine-tuning, we expect you to:
328
+
329
+
* Have a dataset `kontext-community/relighting`
330
+
* Supply `image_column`, `cond_image_column`, and `caption_column` values when launching training
331
+
297
332
### Misc notes
298
333
299
334
* By default, we use `mode` as the value of `--vae_encode_mode` argument. This is because Kontext uses `mode()` of the distribution predicted by the VAE instead of sampling from it.
0 commit comments