File tree Expand file tree Collapse file tree 1 file changed +1
-5
lines changed
vllm/v1/attention/backends Expand file tree Collapse file tree 1 file changed +1
-5
lines changed Original file line number Diff line number Diff line change 4
4
5
5
import ast
6
6
from dataclasses import dataclass
7
- from typing import TYPE_CHECKING , Any , Optional
7
+ from typing import TYPE_CHECKING , Optional
8
8
9
9
import torch
10
10
@@ -313,15 +313,11 @@ def __init__(
313
313
alibi_slopes : Optional [list [float ]],
314
314
sliding_window : Optional [int ],
315
315
kv_cache_dtype : str ,
316
- blocksparse_params : Optional [dict [str , Any ]] = None ,
317
316
logits_soft_cap : Optional [float ] = None ,
318
317
attn_type : AttentionType = AttentionType .DECODER ,
319
318
kv_sharing_target_layer_name : Optional [str ] = None ,
320
319
use_irope : bool = False ,
321
320
) -> None :
322
- if blocksparse_params is not None :
323
- raise ValueError (
324
- "TreeAttention does not support block-sparse attention." )
325
321
self .num_heads = num_heads
326
322
self .head_size = head_size
327
323
self .scale = float (scale )
You can’t perform that action at this time.
0 commit comments