We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 4f41615 commit f8d0820Copy full SHA for f8d0820
tests/kernels/attention/test_flash_attn.py
@@ -145,7 +145,7 @@ def test_flash_attn_with_paged_kv(
145
v_descale = None
146
if q_dtype is not None:
147
# QKV are drawn from N(0, 1): no need for a fp8 scaling factor
148
- maybe_quantized_query = query.to(q_dtype)
+ maybe_quantized_query = q.to(q_dtype)
149
maybe_quantized_key_cache = key_cache.to(q_dtype)
150
maybe_quantized_value_cache = value_cache.to(q_dtype)
151
0 commit comments