Skip to content

Commit 41f43d8

Browse files
committed
fix & remove splitfuse router mode
1 parent 0e459ea commit 41f43d8

File tree

7 files changed

+4
-187
lines changed

7 files changed

+4
-187
lines changed

lightllm/server/core/objs/req.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -287,11 +287,9 @@ def post_init(
287287
class ChunkedPrefillReq(Req):
288288
_pack_ = 4
289289

290-
def post_init(self):
291-
args = get_env_start_args()
292-
self.max_waiting_token = args.router_max_wait_tokens
293-
294290
def get_tuple_tokens(self, is_busy, router_max_new_token_len):
291+
args = get_env_start_args()
292+
max_waiting_token = args.router_max_wait_tokens
295293
has_out_len = self.shm_cur_output_len
296294
if self.sample_params.ignore_eos:
297295
cur_max_new_token_len = self.sample_params.max_new_tokens
@@ -306,7 +304,7 @@ def get_tuple_tokens(self, is_busy, router_max_new_token_len):
306304
b_len = (
307305
(self.input_len + has_out_len - self.shm_cur_kv_len + self.chunked_prefill_size - 1)
308306
// self.chunked_prefill_size
309-
* (self.max_waiting_token + 1)
307+
* (max_waiting_token + 1)
310308
+ cur_max_new_token_len
311309
- has_out_len
312310
- 1

lightllm/server/router/model_infer/mode_backend/__init__.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,6 @@
22
from .continues_batch.impl_for_return_all_prompt_logprobs import ReturnPromptLogProbBackend
33
from .continues_batch.impl_for_reward_model import RewardModelBackend
44
from .chunked_prefill.impl import ChunkedPrefillBackend
5-
from .splitfuse.impl import SplitFuseBackend
65
from .diverse_backend.impl import DiversehBackend
76
from .continues_batch.impl_for_token_healing import TokenHealingBackend
87
from .continues_batch.impl_for_simple_constraint_mode import SimpleConstraintBackend

lightllm/server/router/model_infer/mode_backend/chunked_prefill/impl.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ def post_handel(self, run_reqs: List[InferReq], next_token_ids, next_token_logpr
5454

5555
req_obj.cur_kv_len = len(req_obj.get_chuncked_input_token_ids())
5656
if req_obj.cur_kv_len < req_obj.get_cur_total_len():
57-
return
57+
continue
5858

5959
req_obj.set_next_gen_token_id(next_token_id, next_token_logprob)
6060
req_obj.cur_output_len += 1

lightllm/server/router/model_infer/mode_backend/splitfuse/__init__.py

Whitespace-only changes.

lightllm/server/router/model_infer/mode_backend/splitfuse/impl.py

Lines changed: 0 additions & 93 deletions
This file was deleted.

lightllm/server/router/model_infer/mode_backend/splitfuse/pre_process.py

Lines changed: 0 additions & 83 deletions
This file was deleted.

lightllm/server/router/model_infer/model_rpc.py

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,6 @@
99
from lightllm.server.router.model_infer.mode_backend import (
1010
ContinuesBatchBackend,
1111
ReturnPromptLogProbBackend,
12-
SplitFuseBackend,
1312
ChunkedPrefillBackend,
1413
DiversehBackend,
1514
RewardModelBackend,
@@ -100,7 +99,6 @@ def init_model(self, kvargs):
10099
# 填充真正的 rank_id 参数
101100
kvargs["rank_id"] = self.tp_rank
102101
self.world_size = kvargs["world_size"]
103-
is_splitfuse_mode = kvargs.get("is_splitfuse_mode", False)
104102
enable_chunked_prefill = kvargs.get("enable_chunked_prefill", False)
105103
return_all_prompt_logprobs = kvargs.get("return_all_prompt_logprobs", False)
106104
use_reward_model = kvargs.get("use_reward_model", False)
@@ -124,8 +122,6 @@ def init_model(self, kvargs):
124122
self.backend = ChunkedPrefillBackend()
125123
elif use_reward_model:
126124
self.backend = RewardModelBackend()
127-
elif is_splitfuse_mode:
128-
self.backend = SplitFuseBackend()
129125
elif return_all_prompt_logprobs:
130126
self.backend = ReturnPromptLogProbBackend()
131127
elif diverse_mode:

0 commit comments

Comments
 (0)