Skip to content

Commit 743cfb4

Browse files
committed
fix patch bug
Signed-off-by: wangxiyuan <[email protected]>
1 parent a92fe2a commit 743cfb4

File tree

1 file changed

+1
-0
lines changed

1 file changed

+1
-0
lines changed

vllm_ascend/patch/worker/patch_common/patch_attention_layer.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -196,6 +196,7 @@ def __init__(
196196
self.q_range = torch.tensor(envs.Q_SCALE_CONSTANT, dtype=torch.float32)
197197
self.k_range = torch.tensor(envs.K_SCALE_CONSTANT, dtype=torch.float32)
198198
self.v_range = torch.tensor(envs.V_SCALE_CONSTANT, dtype=torch.float32)
199+
self.query_quant = None
199200

200201

201202
vllm.attention.Attention = AscendAttention

0 commit comments

Comments
 (0)