We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
2 parents f9ca98a + f86feed commit 39db419Copy full SHA for 39db419
fms_mo/aiu_addons/fp8/fp8_attn.py
@@ -29,6 +29,7 @@
29
# Third Party
30
from fms.modules.attention import (
31
AttentionKwargs,
32
+ _sdpa_compute_op,
33
_sdpa_update_attn_kwargs,
34
register_attention_op,
35
)
@@ -340,7 +341,7 @@ def __spyre_scaled_paged_validate_attn_kwargs_op(
340
341
register_attention_op(
342
"spyre_paged_attn_fp8",
343
_spyre_scaled_paged_store_op,
- compute_op=_math_fp8_compute_op,
344
+ compute_op=_sdpa_compute_op,
345
is_prefill_op=lambda **attn_kwargs: attn_kwargs.get("block_table", None)
346
is None,
347
compute_decode_op=_spyre_scaled_paged_compute_op,
0 commit comments