From 785c416fde9e35e09e817676dc6fd38b49bf40fe Mon Sep 17 00:00:00 2001 From: root Date: Sun, 28 Sep 2025 09:34:15 +0000 Subject: [PATCH] [attn](fix): fix attention bug of position assert --- vllm/attention/layer.py | 1 - 1 file changed, 1 deletion(-) diff --git a/vllm/attention/layer.py b/vllm/attention/layer.py index 3697275a7b64..5e455a2852f4 100644 --- a/vllm/attention/layer.py +++ b/vllm/attention/layer.py @@ -541,7 +541,6 @@ def unified_attention_with_output( output_scale=output_scale, positions=positions) else: - assert positions is None, f"positions must be None {positions=}" self.impl.forward(self, query, key,