We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent a1edc02 commit a9ed330Copy full SHA for a9ed330
lightllm/common/fused_moe/grouped_fused_moe.py
@@ -1000,13 +1000,13 @@ def outplace_fused_experts_impl_fake(
1000
hidden_states: torch.Tensor,
1001
w1: torch.Tensor,
1002
w2: torch.Tensor,
1003
- # optional bias for w1 and w2
1004
- w1_bias: Optional[torch.Tensor],
1005
- w2_bias: Optional[torch.Tensor],
1006
topk_weights: torch.Tensor,
1007
topk_ids: torch.Tensor,
1008
use_fp8_w8a8: bool = False,
1009
use_int8_w8a16: bool = False,
+ # optional bias for w1 and w2
+ w1_bias: Optional[torch.Tensor] = None,
+ w2_bias: Optional[torch.Tensor] = None,
1010
w1_scale: Optional[torch.Tensor] = None,
1011
w2_scale: Optional[torch.Tensor] = None,
1012
a1_scale: Optional[torch.Tensor] = None,
0 commit comments