Skip to content

Commit 3611045

Browse files
authored
support reasoning_max_tokens (#4207)
1 parent 0bfffdb commit 3611045

File tree

1 file changed

+2
-4
lines changed

1 file changed

+2
-4
lines changed

fastdeploy/entrypoints/engine_client.py

Lines changed: 2 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -149,8 +149,6 @@ async def add_requests(self, task):
149149
task["prompt_token_ids_len"] = len(task["prompt_token_ids"])
150150
input_ids_len = task["prompt_token_ids_len"]
151151
task["max_tokens"] = min(self.max_model_len - input_ids_len, task.get("max_tokens"))
152-
if task.get("reasoning_max_tokens", None) is None:
153-
task["reasoning_max_tokens"] = max(int(task["max_tokens"] * 0.8), 1)
154152
min_tokens = task.get("min_tokens", 1)
155153
if "messages" in task:
156154
del task["messages"]
@@ -229,8 +227,8 @@ def vaild_parameters(self, data):
229227
raise ValueError(f"max_tokens can be defined [1, {self.max_model_len}).")
230228

231229
if data.get("reasoning_max_tokens") is not None:
232-
if data["reasoning_max_tokens"] > data["max_tokens"] or data["reasoning_max_tokens"] < 1:
233-
raise ValueError("reasoning_max_tokens must be between max_tokens and 1")
230+
if data["reasoning_max_tokens"] > data["max_tokens"] or data["reasoning_max_tokens"] < 0:
231+
raise ValueError("reasoning_max_tokens must be between max_tokens and 0")
234232

235233
if data.get("top_p") is not None:
236234
if data["top_p"] > 1 or data["top_p"] < 0:

0 commit comments

Comments
 (0)