Skip to content

Commit bfdec9f

Browse files
committed
fix
1 parent c433e05 commit bfdec9f

File tree

3 files changed

+6
-11
lines changed

3 files changed

+6
-11
lines changed

fastdeploy/engine/request.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -72,7 +72,6 @@ def __init__(
7272
structural_tag: Optional[Any] = None,
7373
guided_json_object: Optional[bool] = None,
7474
enable_thinking: Optional[bool] = False,
75-
model_status: Optional[str] = None,
7675
trace_carrier: dict = dict(),
7776
dp_rank: Optional[int] = None,
7877
chat_template: Optional[str] = None,

fastdeploy/input/ernie4_5_processor.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -156,11 +156,9 @@ def process_request(self, request, max_model_len=None, **kwargs):
156156
if self.reasoning_parser and self.reasoning_parser.__class__.__name__ == "ErnieX1ReasoningParser":
157157
request.enable_thinking = True
158158
if self.reasoning_parser:
159-
self.model_status_dict[request.request_id] = self.reasoning_parser.get_model_status(
160-
request.prompt_token_ids
161-
)
162-
if self.model_status_dict[request.request_id] == "think_start":
163-
request.enable_thinking = True
159+
model_status = self.reasoning_parser.get_model_status(request.prompt_token_ids)
160+
self.model_status_dict[request.request_id] = model_status
161+
request.enable_thinking = model_status == "think_start"
164162

165163
data_processor_logger.info(f"Processed request: {request}")
166164
return request

fastdeploy/input/text_processor.py

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -267,11 +267,9 @@ def process_request(self, request, max_model_len=None, **kwargs):
267267
if request.get("top_p") < _SAMPLING_EPS:
268268
request.set("top_p", _SAMPLING_EPS)
269269
if self.reasoning_parser:
270-
self.model_status_dict[request.request_id] = self.reasoning_parser.get_model_status(
271-
request.prompt_token_ids
272-
)
273-
if self.model_status_dict[request.request_id] == "think_start":
274-
request.enable_thinking = True
270+
model_status = self.reasoning_parser.get_model_status(request.prompt_token_ids)
271+
self.model_status_dict[request.request_id] = model_status
272+
request.enable_thinking = model_status == "think_start"
275273

276274
data_processor_logger.info(f"Processed request: {request}")
277275
return request

0 commit comments

Comments
 (0)