Skip to content

Commit ec25768

Browse files
Lay2000aoyuli
andauthored
fix: support use_sync flag in xFuserLongContextAttention (#505)
Co-authored-by: aoyuli <aoyuli@bytedance.com>
1 parent a440a6e commit ec25768

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

xfuser/core/long_ctx_attention/hybrid/attn_layer.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -29,6 +29,7 @@ def __init__(
2929
ring_impl_type: str = "basic",
3030
use_pack_qkv: bool = False,
3131
use_kv_cache: bool = False,
32+
use_sync: bool = False,
3233
attn_type: AttnType = AttnType.FA,
3334
attn_processor: torch.nn.Module = None,
3435
) -> None:
@@ -47,6 +48,7 @@ def __init__(
4748
gather_idx=gather_idx,
4849
ring_impl_type=ring_impl_type,
4950
use_pack_qkv=use_pack_qkv,
51+
use_sync=use_sync,
5052
attn_type = attn_type,
5153
)
5254
self.use_kv_cache = use_kv_cache

0 commit comments

Comments
 (0)