You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: tensorrt_llm/executor/result.py
+12Lines changed: 12 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -102,6 +102,8 @@ class CompletionOutput:
102
102
finish_reason (Literal['stop', 'length', 'timeout', 'cancelled'], optional): The reason why the sequence is finished. Defaults to None.
103
103
stop_reason (int, str, optional): The stop string or token id that caused the completion to stop, None if the completion finished for some other reason. Defaults to None.
104
104
generation_logits (torch.Tensor, optional): The logits on the generated output token ids. Defaults to None.
105
+
additional_context_outputs (Dict[str, torch.Tensor], optional): The additional context outputs. Defaults to None.
106
+
additional_generation_outputs (Dict[str, torch.Tensor], optional): The additional generation outputs. Defaults to None.
105
107
disaggregated_params (tensorrt_llm.disaggregated_params.DisaggregatedParams, optional): Parameters needed for disaggregated serving. Includes the type of request, the first generated tokens, the context request id and the any additional state needing to be transferred from context and generation instances. Defaults to None.
106
108
request_perf_metrics (tensorrt_llm.bindings.executor.RequestPerfMetrics, optional): Performance metrics for the request. Defaults to None.
0 commit comments