We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4d4297e commit cbc8457Copy full SHA for cbc8457
vllm/model_executor/models/qwen2_5_vl.py
@@ -396,13 +396,13 @@ def forward(
396
max_seqlen: Optional[int] = None, # Only used for Flash Attention
397
seqlens: Optional[list[int]] = None, # Only used for xFormers
398
) -> torch.Tensor:
399
- x = x + self.attn(self.norm1(x),
400
- cu_seqlens=cu_seqlens,
401
- rotary_pos_emb=rotary_pos_emb,
402
- max_seqlen=max_seqlen,
403
- seqlens=seqlens)
404
-
405
- x = x + self.mlp(self.norm2(x))
+ x_attn = self.attn(self.norm1(x),
+ cu_seqlens=cu_seqlens,
+ rotary_pos_emb=rotary_pos_emb,
+ max_seqlen=max_seqlen,
+ seqlens=seqlens)
+ x_fused_norm, residual = self.norm2(x, residual=x_attn)
+ x = residual + self.mlp(x_fused_norm)
406
return x
407
408
0 commit comments