You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
- Add include_prompt_progress parameter to slot_params (default: false)
- Extend server_task_result_cmpl_partial with progress fields
- Implement send_progress_response() function with 1% progress intervals
- Add progress response in prompt processing loop
- Update JSON response to include prompt_processing field when requested
- Add comprehensive documentation to README.md
- Ensure full backward compatibility with existing clients
Closes#14685
Copy file name to clipboardExpand all lines: tools/server/README.md
+2Lines changed: 2 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -428,6 +428,8 @@ By default, this value is set to `0`, meaning no tokens are kept. Use `-1` to re
428
428
429
429
`stream`: Allows receiving each predicted token in real-time instead of waiting for the completion to finish (uses a different response format). To enable this, set to `true`.
430
430
431
+
`include_prompt_progress`: When `stream` is enabled, this option allows receiving prompt processing progress information before the text generation begins. The progress responses contain a `prompt_processing` field with details about the number of tokens processed and overall progress. This is useful for long prompts where users want to see evaluation progress instead of waiting silently. Default: `false` (only applies when `stream` is `true`).
432
+
431
433
`stop`: Specify a JSON array of stopping strings.
432
434
These words will not be included in the completion, so make sure to add them to the prompt for the next iteration. Default: `[]`
0 commit comments