About using streaming while waiting for model response #5968
miguellobato84
started this conversation in
Feature Requests & Suggestions
Replies: 2 comments 1 reply
-
Maybe they would make sense as a scheduled task that runs in the background? That feature is on the new roadmap for the year: |
Beta Was this translation helpful? Give feedback.
1 reply
-
I feel like this issue became more apparent recently, with GPT-5 being super slow¹. Multiple people reported issues on Discord. Might be worth considering to prioritize work on scheduled tasks? ¹) As per Artifical Analysis: "GPT-5 (high) has a higher latency compared to average, taking 79.21s to receive the first token (TTFT)." |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi, I'm wondering if makes sense to use a streaming (aka keep waiting for) the response for the model. For some models like 4o it makes sense, but for others like o3 I don't think so and we should change to a poll approach
The problem arises when the app is behind some kind of firewall that closes connections without activity after 10-30s, so models like o3 becomes unusable.
Problems will also appear when research models, which can take upto 30 mins, are added. We can´t keep an active connection without receiving data for that amount of time.
Beta Was this translation helpful? Give feedback.
All reactions