Skip to content

Commit c6e2a37

Browse files
authored
[BugFix] qwen2.5vl enable_thinking=true bug fix (#3920)
1 parent 8d77c1c commit c6e2a37

File tree

1 file changed

+10
-0
lines changed

1 file changed

+10
-0
lines changed

fastdeploy/input/qwen_vl_processor/qwen_vl_processor.py

Lines changed: 10 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -249,6 +249,16 @@ def process_request_dict(self, request, max_model_len=None):
249249
# Handle continuation of previous generation by appending existing tokens
250250
if metadata and metadata.get("generated_token_ids"):
251251
self.append_generated_tokens(outputs, metadata["generated_token_ids"])
252+
253+
enable_thinking = False
254+
if metadata:
255+
enable_thinking = metadata.get("enable_thinking", False)
256+
257+
if request.get("chat_template_kwargs"):
258+
chat_template_kwargs = request.get("chat_template_kwargs")
259+
enable_thinking = chat_template_kwargs.get("enable_thinking", False)
260+
request["enable_thinking"] = enable_thinking
261+
252262
outputs = self.pack_outputs(outputs)
253263

254264
request["prompt_token_ids"] = outputs["input_ids"].tolist()

0 commit comments

Comments
 (0)