Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion src/open_responses_server/api_controller.py
Original file line number Diff line number Diff line change
Expand Up @@ -403,7 +403,7 @@ async def stream_proxy():
async with client.stream(request.method, url, headers=headers, content=body, timeout=120.0) as response:
async for chunk in response.aiter_bytes():
yield chunk
return StreamingResponse(stream_proxy(), media_type=response.headers.get("content-type"))
return StreamingResponse(stream_proxy(), media_type=request.headers.get('accept', 'application/json'))
else:
response = await client.request(request.method, url, headers=headers, content=body, timeout=120.0)
return Response(content=response.content, status_code=response.status_code, headers=response.headers)
Expand Down
4 changes: 2 additions & 2 deletions test_tools/config.sh
Original file line number Diff line number Diff line change
Expand Up @@ -3,9 +3,9 @@
# This file should be sourced by other scripts

# Default model ID to use if DEFAULT_MODEL_ID environment variable is not set
DEFAULT_MODEL_FALLBACK="meta-llama/llama-4-scout-17b-16e-instruct"
DEFAULT_MODEL_FALLBACK="gpt-4o"
# For codex script which uses a different default model
DEFAULT_CODEX_MODEL_FALLBACK="meta-llama/llama-4-maverick-17b-128e-instruct"
DEFAULT_CODEX_MODEL_FALLBACK="gpt-4o"

if [ "$USE_CODEX" = "true" ]; then
MODEL_ID=${DEFAULT_MODEL_ID:-"$DEFAULT_CODEX_MODEL_FALLBACK"}
Expand Down
Loading