We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent a440a6e commit ec25768Copy full SHA for ec25768
xfuser/core/long_ctx_attention/hybrid/attn_layer.py
@@ -29,6 +29,7 @@ def __init__(
29
ring_impl_type: str = "basic",
30
use_pack_qkv: bool = False,
31
use_kv_cache: bool = False,
32
+ use_sync: bool = False,
33
attn_type: AttnType = AttnType.FA,
34
attn_processor: torch.nn.Module = None,
35
) -> None:
@@ -47,6 +48,7 @@ def __init__(
47
48
gather_idx=gather_idx,
49
ring_impl_type=ring_impl_type,
50
use_pack_qkv=use_pack_qkv,
51
+ use_sync=use_sync,
52
attn_type = attn_type,
53
)
54
self.use_kv_cache = use_kv_cache
0 commit comments