Skip to content

Commit 771ffeb

Browse files
committed
fix
1 parent 7cf5fdb commit 771ffeb

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

lightllm/models/qwen3next/mem_manager.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -104,6 +104,12 @@ def free_all(self):
104104
self.ssm_state_mem_manager.free_all()
105105
return
106106

107+
@override
108+
def get_cell_size(self):
109+
# Only full attention layers and MTP layers have KV cache
110+
kv_cache_layer_num = self.full_attn_layer_num + self.mtp_layer_num
111+
return 2 * self.head_num * self.head_dim * kv_cache_layer_num * torch._utils._element_size(self.dtype)
112+
107113
@override
108114
def get_buffer(self, layer_index) -> Tuple[torch.Tensor, torch.Tensor]:
109115
assert layer_index < self.layer_num, "layer_index is out of range"

0 commit comments

Comments
 (0)