You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
feat(vlm): track generated tokens and stop reasons for VLM models (#2543)
* feat: add enum StopReason and use it in VlmPrediction
Signed-off-by: ElHachem02 <[email protected]>
* add vlm_inference time for api calls and track stop reason
Signed-off-by: ElHachem02 <[email protected]>
* fix: rename enum to VlmStopReason
Signed-off-by: ElHachem02 <[email protected]>
* Propagate partial success status if page reaches max tokens
Signed-off-by: ElHachem02 <[email protected]>
* feat: page with generation stopped by loop detector create partial success status
Signed-off-by: Peter El Hachem <[email protected]>
* Add hint for future improvement
Signed-off-by: Peter El Hachem <[email protected]>
* fix: remove vlm_stop_reason from extracted page data, add UNSPECIFIED state as VlmStopReason to avoid null value
Signed-off-by: Peter El Hachem <[email protected]>
---------
Signed-off-by: ElHachem02 <[email protected]>
Signed-off-by: Peter El Hachem <[email protected]>
Co-authored-by: Peter El Hachem <[email protected]>
0 commit comments