Skip to content

Commit cbc8457

Browse files
vllmellmkliuae
andauthored
[Model] Switch to Fused RMS norm in Qwen2.5_VL model. (#22184)
Signed-off-by: kf <[email protected]> Signed-off-by: tjtanaavllm <[email protected]> Signed-off-by: vllmellm <[email protected]> Co-authored-by: kf <[email protected]>
1 parent 4d4297e commit cbc8457

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

vllm/model_executor/models/qwen2_5_vl.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -396,13 +396,13 @@ def forward(
396396
max_seqlen: Optional[int] = None, # Only used for Flash Attention
397397
seqlens: Optional[list[int]] = None, # Only used for xFormers
398398
) -> torch.Tensor:
399-
x = x + self.attn(self.norm1(x),
400-
cu_seqlens=cu_seqlens,
401-
rotary_pos_emb=rotary_pos_emb,
402-
max_seqlen=max_seqlen,
403-
seqlens=seqlens)
404-
405-
x = x + self.mlp(self.norm2(x))
399+
x_attn = self.attn(self.norm1(x),
400+
cu_seqlens=cu_seqlens,
401+
rotary_pos_emb=rotary_pos_emb,
402+
max_seqlen=max_seqlen,
403+
seqlens=seqlens)
404+
x_fused_norm, residual = self.norm2(x, residual=x_attn)
405+
x = residual + self.mlp(x_fused_norm)
406406
return x
407407

408408

0 commit comments

Comments
 (0)