We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent bcc4e69 commit 25e4e87Copy full SHA for 25e4e87
vllm/model_executor/layers/utils.py
@@ -94,16 +94,6 @@ def default_unquantized_gemm(layer: torch.nn.Module,
94
bias: Optional[torch.Tensor] = None):
95
return torch.nn.functional.linear(x, weight, bias)
96
97
-def aiter_GEMM_check(m, n, k):
98
- if ((n == 5120 and k == 2880)
99
- or (n == 2880 and k == 4096)
100
- or (n == 128 and k == 2880)
101
- or (n == 640 and k == 2880)
102
- or (n == 2880 and k == 512)):
103
- return True
104
- return False
105
-
106
107
def aiter_GEMM_check(m, n, k):
108
if ((n == 5120 and k == 2880)
109
or (n == 2880 and k == 4096)
0 commit comments