7
7
8
8
import os
9
9
from dataclasses import dataclass
10
- from typing import Any , Optional
10
+ from typing import Optional
11
11
12
12
import torch
13
13
import vllm_gaudi .extension .kernels as kernels
@@ -161,7 +161,6 @@ def __init__(
161
161
alibi_slopes : Optional [list [float ]],
162
162
sliding_window : Optional [int ],
163
163
kv_cache_dtype : str ,
164
- blocksparse_params : Optional [dict [str , Any ]],
165
164
logits_soft_cap : Optional [float ],
166
165
attn_type : str ,
167
166
kv_sharing_target_layer_name : Optional [str ] = None ,
@@ -170,7 +169,7 @@ def __init__(
170
169
torch .nn .Module .__init__ (self )
171
170
MLACommonImpl .__init__ (self , num_heads , head_size , scale , num_kv_heads ,
172
171
alibi_slopes , sliding_window , kv_cache_dtype ,
173
- blocksparse_params , logits_soft_cap , attn_type ,
172
+ logits_soft_cap , attn_type ,
174
173
kv_sharing_target_layer_name , ** kwargs )
175
174
self .enable_fp8_attn = kv_cache_dtype == 'fp8_inc' and os .environ .get (
176
175
'QUANT_CONFIG' , None ) is None
@@ -191,13 +190,11 @@ def __init__(
191
190
assert self .prefill_impl != 'fsdpa_impl' or alibi_slopes is None , \
192
191
'Prefill with FusedSDPA not supported with alibi slopes!'
193
192
194
- unsupported_features = [
195
- alibi_slopes , sliding_window , blocksparse_params , logits_soft_cap
196
- ]
193
+ unsupported_features = [alibi_slopes , sliding_window , logits_soft_cap ]
197
194
if any (unsupported_features ):
198
195
raise NotImplementedError (
199
196
"HPUMLAImpl does not support one of the following: "
200
- "alibi_slopes, sliding_window, blocksparse_params, "
197
+ "alibi_slopes, sliding_window, "
201
198
"logits_soft_cap" )
202
199
203
200
if attn_type != AttentionType .DECODER :
@@ -379,7 +376,6 @@ def __init__(
379
376
alibi_slopes : Optional [list [float ]],
380
377
sliding_window : Optional [int ],
381
378
kv_cache_dtype : str ,
382
- blocksparse_params : Optional [dict [str , Any ]] = None ,
383
379
logits_soft_cap : Optional [float ] = None ,
384
380
attn_type : str = AttentionType .DECODER ,
385
381
kv_sharing_target_layer_name : Optional [str ] = None ,
0 commit comments