Skip to content

Commit bfe0b20

Browse files
committed
Set enable_attn_fusion=True by default now that attention fusion is fixed
1 parent 55cd9df commit bfe0b20

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

vllm/config/compilation.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,7 @@ class PassConfig:
7777

7878
enable_fusion: bool = field(default_factory=lambda: envs.VLLM_USE_V1)
7979
"""Whether to enable the custom fusion (RMSNorm/SiluMul+quant) pass."""
80-
enable_attn_fusion: bool = False
80+
enable_attn_fusion: bool = field(default_factory=lambda: envs.VLLM_USE_V1)
8181
"""Whether to enable the custom attention+quant fusion pass."""
8282
enable_noop: bool = field(default_factory=lambda: envs.VLLM_USE_V1)
8383
"""Whether to enable the custom no-op elimination pass."""

0 commit comments

Comments
 (0)