2323)
2424from vllm .entrypoints .openai .engine .protocol import ErrorResponse
2525from vllm .logprobs import Logprob
26- from vllm .transformers_utils . tokenizer import AnyTokenizer
26+ from vllm .tokenizers import TokenizerLike
2727
2828from djl_python .outputs import Output
2929from djl_python .async_utils import create_non_stream_output , create_stream_chunk_output
@@ -101,7 +101,7 @@ def convert_lmi_schema_to_completion_request(
101101
102102def convert_completion_logprobs_to_tgi_tokens (
103103 completion_logprobs : CompletionLogProbs ,
104- tokenizer : AnyTokenizer ,
104+ tokenizer : TokenizerLike ,
105105) -> List [dict ]:
106106 token_logprobs = completion_logprobs .token_logprobs
107107 tokens = completion_logprobs .tokens
@@ -138,7 +138,7 @@ def convert_completion_response_to_lmi_schema(
138138 response : CompletionResponse ,
139139 request : CompletionRequest = None ,
140140 include_details : bool = False ,
141- tokenizer : AnyTokenizer = None ,
141+ tokenizer : TokenizerLike = None ,
142142) -> Output :
143143 primary_choice = response .choices [0 ]
144144 lmi_response = {"generated_text" : primary_choice .text }
@@ -258,7 +258,7 @@ def convert_completion_chunk_response_to_lmi_schema(
258258def lmi_with_details_non_stream_output_formatter (
259259 response : CompletionResponse ,
260260 request : CompletionRequest = None ,
261- tokenizer : AnyTokenizer = None ,
261+ tokenizer : TokenizerLike = None ,
262262) -> Output :
263263 return convert_completion_response_to_lmi_schema (response ,
264264 include_details = True ,
@@ -269,7 +269,7 @@ def lmi_with_details_non_stream_output_formatter(
269269def lmi_non_stream_output_formatter (
270270 response : CompletionResponse ,
271271 request : CompletionRequest = None ,
272- tokenizer : AnyTokenizer = None ,
272+ tokenizer : TokenizerLike = None ,
273273) -> Output :
274274 return convert_completion_response_to_lmi_schema (response ,
275275 include_details = False ,
0 commit comments