Skip to content

Commit 15ed4a3

Browse files
Merge pull request #764 from yasinBursali/fix/chat-thinking-model-max-tokens
fix: handle thinking model responses in /api/chat endpoint
2 parents e55fc4e + 14c4bee commit 15ed4a3

File tree

1 file changed

+4
-1
lines changed
  • dream-server/extensions/services/dashboard-api/routers

1 file changed

+4
-1
lines changed

dream-server/extensions/services/dashboard-api/routers/setup.py

Lines changed: 4 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@
44
import json
55
import logging
66
import os
7+
import re
78
from datetime import datetime, timezone
89
from pathlib import Path
910

@@ -170,7 +171,7 @@ async def chat(request: ChatRequest, api_key: str = Depends(verify_api_key)):
170171
payload = {
171172
"model": model,
172173
"messages": [{"role": "system", "content": system_prompt}, {"role": "user", "content": request.message}],
173-
"max_tokens": 256, "temperature": 0.7
174+
"max_tokens": 2048, "temperature": 0.7
174175
}
175176

176177
try:
@@ -180,6 +181,8 @@ async def chat(request: ChatRequest, api_key: str = Depends(verify_api_key)):
180181
if resp.status == 200:
181182
data = await resp.json()
182183
response_text = data.get("choices", [{}])[0].get("message", {}).get("content", "")
184+
# Strip thinking model tags — content may contain <think>...</think> blocks
185+
response_text = re.sub(r'<think>[\s\S]*?</think>\s*', '', response_text).strip()
183186
return {"response": response_text, "success": True}
184187
else:
185188
error_text = await resp.text()

0 commit comments

Comments
 (0)