23
23
log_tracing_disabled_warning ,
24
24
)
25
25
from vllm .transformers_utils .tokenizer import AnyTokenizer # noqa: TCH002
26
- from vllm .transformers_utils .tokenizers .mistral import MistralTokenizer
27
26
from vllm .utils import iterate_with_cancellation
28
27
29
28
from vllm_tgis_adapter .logging import init_logger
@@ -856,12 +855,16 @@ async def Tokenize(
856
855
tokenizer = await self ._get_tokenizer (adapter_kwargs )
857
856
858
857
responses : list [TokenizeResponse ] = []
859
- is_mistral_tokenizer = isinstance (tokenizer , MistralTokenizer )
860
858
861
859
# TODO: maybe parallelize, also move convert_ids_to_tokens into the
862
860
# other threads
863
861
for req in request .requests :
864
- if is_mistral_tokenizer :
862
+ if not hasattr (tokenizer , "encode_plus" ):
863
+ if request .return_offsets :
864
+ raise ValueError (
865
+ f"{ type (tokenizer )} doesn't support "
866
+ "return_offsets at the moment. "
867
+ )
865
868
token_ids = tokenizer .encode (
866
869
prompt = req .text ,
867
870
)
@@ -885,11 +888,6 @@ async def Tokenize(
885
888
offsets = None
886
889
887
890
if request .return_offsets :
888
- if is_mistral_tokenizer :
889
- raise ValueError (
890
- "Mistral tokenizer doesn't support "
891
- "return_offsets at the moment. "
892
- )
893
891
offsets = [
894
892
{"start" : start , "end" : end }
895
893
for start , end in batch_encoding .offset_mapping
0 commit comments