Skip to content

Commit 087ca04

Browse files
author
niushengxiao
committed
fix: fix a bug in flashinfer
1 parent a4bc0d6 commit 087ca04

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

lightllm/models/llama/flashinfer_struct.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -81,8 +81,8 @@ def init_some_extra_state(self, model, input_ids: torch.Tensor):
8181
self.req_manager.req_to_token_indexs,
8282
self.b_req_idx,
8383
self.b_seq_len,
84-
kv_starts,
85-
self.max_len_in_batch,
84+
kv_starts[:-1],
85+
self.max_kv_seq_len,
8686
kv_indices,
8787
)
8888
self.prefill_wrapper = flashinfer.prefill.BatchPrefillWithPagedKVCacheWrapper(

0 commit comments

Comments
 (0)