Skip to content

Commit c3d5e61

Browse files
author
liujiacheng
committed
rename
1 parent 088345b commit c3d5e61

File tree

2 files changed

+6
-6
lines changed

2 files changed

+6
-6
lines changed

lightllm/server/router/model_infer/mode_backend/base_backend.py

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
from lightllm.server.router.shm_reqs_io_buffer import ShmReqsIOBuffer
3232
from lightllm.server.router.model_infer.mode_backend.overlap_events import OverlapEventManager, OverlapEventPack
3333
from lightllm.models.deepseek_mtp.model import Deepseek3MTPModel
34-
from .multi_level_cache_manager import MultiLevelCacheManager
34+
from .multi_level_cache import MultiLevelCacheModule
3535

3636

3737
class ModeBackend:
@@ -201,7 +201,7 @@ def init_model(self, kvargs):
201201
self.infer_loop_thread1.start()
202202

203203
if self.args.enable_cpu_cache:
204-
self.multi_level_cache_manager = MultiLevelCacheManager(self)
204+
self.multi_level_cache_module = MultiLevelCacheModule(self)
205205
return
206206

207207
def init_custom(self):
@@ -355,7 +355,7 @@ def _init_reqs(self, reqs: List[Tuple]):
355355
def _fill_cpu_cache_to_reqs(self, req_ids):
356356
req_objs: List[InferReq] = [g_infer_context.requests_mapping[req_id] for req_id in req_ids]
357357
g_infer_state_lock.acquire()
358-
self.multi_level_cache_manager.fill_cpu_cache_to_reqs(reqs=req_objs)
358+
self.multi_level_cache_module.fill_cpu_cache_to_reqs(reqs=req_objs)
359359
g_infer_state_lock.release()
360360
return
361361

@@ -386,7 +386,7 @@ def _get_classed_reqs(
386386
5. decode_reqs 需要进行decode操作的请求
387387
"""
388388
if self.args.enable_cpu_cache:
389-
self.multi_level_cache_manager.update_cpu_cache_task_states()
389+
self.multi_level_cache_module.update_cpu_cache_task_states()
390390

391391
if req_ids is None:
392392
req_ids = g_infer_context.infer_req_ids
@@ -469,7 +469,7 @@ def _get_classed_reqs(
469469
self._pre_handle_finished_reqs(finished_reqs=finished_reqs)
470470
# 如果使能了 cpu cache 功能,对于已经完成的请求,进行 gpu kv 卸载到 cpu cache的操作。
471471
if self.args.enable_cpu_cache:
472-
true_finished_reqs = self.multi_level_cache_manager.handle_finished_reqs(finished_reqs=finished_reqs)
472+
true_finished_reqs = self.multi_level_cache_module.handle_finished_reqs(finished_reqs=finished_reqs)
473473
else:
474474
true_finished_reqs = finished_reqs
475475

lightllm/server/router/model_infer/mode_backend/multi_level_cache_manager.py renamed to lightllm/server/router/model_infer/mode_backend/multi_level_cache.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@
1212
from lightllm.server.router.model_infer.infer_batch import g_infer_context
1313

1414

15-
class MultiLevelCacheManager(object):
15+
class MultiLevelCacheModule(object):
1616
def __init__(self, backend):
1717
self.args = get_env_start_args()
1818
from .base_backend import ModeBackend

0 commit comments

Comments
 (0)