Skip to content

Commit a200780

Browse files
committed
Don't use biases for input projs and output AdaNorm
1 parent 7acd7da commit a200780

File tree

1 file changed

+5
-3
lines changed

1 file changed

+5
-3
lines changed

src/diffusers/models/transformers/transformer_flux2.py

Lines changed: 5 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -641,8 +641,8 @@ def __init__(
641641
self.single_stream_modulation = Flux2Modulation(self.inner_dim, mod_param_sets=1, bias=False)
642642

643643
# 4. Input projections
644-
self.x_embedder = nn.Linear(in_channels, self.inner_dim)
645-
self.context_embedder = nn.Linear(joint_attention_dim, self.inner_dim)
644+
self.x_embedder = nn.Linear(in_channels, self.inner_dim, bias=False)
645+
self.context_embedder = nn.Linear(joint_attention_dim, self.inner_dim, bias=False)
646646

647647
# 5. Double Stream Transformer Blocks
648648
self.transformer_blocks = nn.ModuleList(
@@ -675,7 +675,9 @@ def __init__(
675675
)
676676

677677
# 7. Output layers
678-
self.norm_out = AdaLayerNormContinuous(self.inner_dim, self.inner_dim, elementwise_affine=False, eps=eps)
678+
self.norm_out = AdaLayerNormContinuous(
679+
self.inner_dim, self.inner_dim, elementwise_affine=False, eps=eps, bias=False
680+
)
679681
self.proj_out = nn.Linear(self.inner_dim, patch_size * patch_size * self.out_channels, bias=False)
680682

681683
self.gradient_checkpointing = False

0 commit comments

Comments
 (0)