Have been chatting with @bghira on our Discord forum about the negative effects of not carefully resizing the training data.
@bghira noticed that the way we resize the images in our training examples can introduce unwanted artifacts in the samples generated by the end fine-tuned model. In short, it ruins the fidelity of the generated samples.
We don't want to make the data input pipelines in our training examples super sophisticated, guaranteeing SoTA results. But that said, I think it's best in the community's interest if we made a note about this phenomenon in the READMEs and just added a comment about it in the pipelines.
Curious to know what @patil-suraj @apolinario think about this. Personally, I think it makes sense.
I will let @bghira share some more insights as well.