Skip to content

Commit 2afd14b

Browse files
committed
refomat
1 parent 296a579 commit 2afd14b

File tree

8 files changed

+43
-17
lines changed

8 files changed

+43
-17
lines changed

lightllm/server/core/objs/req.py

Lines changed: 7 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -95,12 +95,14 @@ class Req(ctypes.Structure):
9595
# 请求回复累计概率和
9696
("cumlogprob", ctypes.c_float),
9797
]
98-
98+
9999
def get_str(self):
100-
return (f"request_id:{self.request_id}, input_len:{self.input_len},"
101-
f"shm_cur_kv_len:{self.shm_cur_kv_len},"
102-
f"shm_cur_output_len:{self.shm_cur_output_len},"
103-
f"finish_status:{self.finish_status.is_finished()}")
100+
return (
101+
f"request_id:{self.request_id}, input_len:{self.input_len},"
102+
f"shm_cur_kv_len:{self.shm_cur_kv_len},"
103+
f"shm_cur_output_len:{self.shm_cur_output_len},"
104+
f"finish_status:{self.finish_status.is_finished()}"
105+
)
104106

105107
def init(
106108
self,

lightllm/server/router/model_infer/model_rpc.py

Lines changed: 30 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -32,9 +32,9 @@ class ModelRpcServer:
3232
def __init__(
3333
self,
3434
args,
35-
tp_rank: int,
36-
local_tp_rank: int,
37-
world_size: int,
35+
tp_rank: int,
36+
local_tp_rank: int,
37+
world_size: int,
3838
local_world_size: int,
3939
rpc_event: multiprocessing.Event,
4040
rpc_finished_event: multiprocessing.Event,
@@ -286,7 +286,9 @@ def _init_env(
286286

287287
g_router_lock.obj = router_lock
288288

289-
model_rpc_server = ModelRpcServer(args, tp_rank, local_tp_rank, world_size, local_world_size, rpc_event, rpc_finished_event, info_queue, mem_queue)
289+
model_rpc_server = ModelRpcServer(
290+
args, tp_rank, local_tp_rank, world_size, local_world_size, rpc_event, rpc_finished_event, info_queue, mem_queue
291+
)
290292
success_event.set()
291293

292294
model_rpc_server.loop_thread.join()
@@ -309,12 +311,34 @@ async def start_model_process(
309311

310312
# 单卡时不使用 rpc
311313
if world_size == 1:
312-
return ModelRpcServer(args, tp_rank, local_tp_rank, world_size, local_world_size, rpc_event, rpc_finished_event, info_queue, mem_queue)
314+
return ModelRpcServer(
315+
args,
316+
tp_rank,
317+
local_tp_rank,
318+
world_size,
319+
local_world_size,
320+
rpc_event,
321+
rpc_finished_event,
322+
info_queue,
323+
mem_queue,
324+
)
313325

314326
success_event = mp.Event()
315327
proc = mp.Process(
316328
target=_init_env,
317-
args=(args, tp_rank, local_tp_rank, world_size, local_world_size, info_queue, mem_queue, router_lock, rpc_event, rpc_finished_event, success_event),
329+
args=(
330+
args,
331+
tp_rank,
332+
local_tp_rank,
333+
world_size,
334+
local_world_size,
335+
info_queue,
336+
mem_queue,
337+
router_lock,
338+
rpc_event,
339+
rpc_finished_event,
340+
success_event,
341+
),
318342
)
319343
proc.start()
320344
success_event.wait(timeout=40)

lightllm/server/router/req_queue/base_queue.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -73,7 +73,7 @@ def get_batch_dp_req_size(self, current_batch: Batch):
7373

7474
return len([req for req in current_batch.reqs if req.sample_params.suggested_dp_index == self.dp_index])
7575

76-
def generate_new_batch(self, current_batch: Batch):
76+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
7777
raise NotImplementedError()
7878

7979
def calcu_batch_token_load(self, current_batch: Batch):

lightllm/server/router/req_queue/chunked_prefill/impl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ def _can_add_new_req(self, req: Req, is_busy, new_batch_first_router_need_tokens
5656
return False, new_batch_first_router_need_tokens
5757

5858
# @calculate_time(show=True, min_cost_ms=10)
59-
def generate_new_batch(self, current_batch: Batch, current_waiting_num):
59+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
6060

6161
# 如果当前已经被调度的请求数量超过了上限,直接不调度新的请求了。
6262
exist_req_num = self.get_batch_dp_req_size(current_batch) + len(self.pause_req_dict)

lightllm/server/router/req_queue/continues_batch/beam_impl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -76,7 +76,7 @@ def _can_add_new_group_reqs(self, cur_handle_group_reqs: List[Req], is_busy, new
7676
return False, new_batch_first_router_need_tokens
7777

7878
# @calculate_time(show=True, min_cost_ms=10)
79-
def generate_new_batch(self, current_batch: Batch):
79+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
8080
# 如果当前已经被调度的请求数量超过了上限,直接不调度新的请求了。
8181
exist_req_num = self.get_batch_dp_req_size(current_batch) + len(self.pause_req_dict)
8282
req_is_full = exist_req_num >= self.running_max_req_size

lightllm/server/router/req_queue/continues_batch/impl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -61,7 +61,7 @@ def _can_add_new_req(self, req: Req, is_busy, new_batch_first_router_need_tokens
6161
return False, new_batch_first_router_need_tokens
6262

6363
# @calculate_time(show=True, min_cost_ms=10)
64-
def generate_new_batch(self, current_batch: Batch, current_waiting_num):
64+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
6565
# 如果当前已经被调度的请求数量超过了上限,直接不调度新的请求了。
6666
exist_req_num = self.get_batch_dp_req_size(current_batch) + len(self.pause_req_dict)
6767
req_is_full = exist_req_num >= self.running_max_req_size

lightllm/server/router/req_queue/continues_batch/pd_decode_impl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ def _init_cache_list(self, current_batch: Batch, is_busy):
2424
return
2525

2626
# @calculate_time(show=True, min_cost_ms=10)
27-
def generate_new_batch(self, current_batch: Batch):
27+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
2828
# 如果当前已经被调度的请求数量超过了上限,直接不调度新的请求了。
2929
exist_req_num = self.get_batch_dp_req_size(current_batch) + len(self.pause_req_dict)
3030
req_is_full = exist_req_num >= self.running_max_req_size

lightllm/server/router/req_queue/dp_base_queue.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ def get_wait_req_num(self):
2727
return sum(queue.get_wait_req_num() for queue in self.inner_queues)
2828

2929
# @calculate_time(show=True, min_cost_ms=10)
30-
def generate_new_batch(self, current_batch: Batch):
30+
def generate_new_batch(self, current_batch: Batch, current_waiting_num: int = -1):
3131
batches = [self.inner_queues[dp_index].generate_new_batch(current_batch) for dp_index in range(self.dp_size)]
3232
return self._merge_batch(batches)
3333

0 commit comments

Comments
 (0)