vllm部署异常 #28
Answered
by
JJJYmmm
SunnyLee20230523
asked this question in
Q&A
vllm部署异常
#28
-
|
我的环境: Package Version
---------------------------------------- -------------
accelerate 1.12.0
addict 2.4.0
aiohappyeyeballs 2.6.1
aiohttp 3.13.3
aiosignal 1.4.0
airportsdata 20260208
annotated-doc 0.0.4
annotated-types 0.7.0
anthropic 0.84.0
anyio 4.12.1
apache-tvm-ffi 0.1.8.post2
astor 0.8.1
asttokens 3.0.1
attrs 25.4.0
av 16.1.0
blake3 1.0.8
blobfile 3.0.0
build 1.4.0
cache-dit 1.2.0
cachetools 7.0.1
cbor2 5.8.0
certifi 2026.2.25
cffi 2.0.0
charset-normalizer 3.4.4
click 8.3.1
cloudpickle 3.1.2
compressed-tensors 0.13.0
cryptography 46.0.5
cuda-bindings 12.9.5
cuda-pathfinder 1.3.5
cuda-python 12.9.0
cupy-cuda12x 14.0.1
datasets 4.5.0
decorator 5.2.1
decord2 3.0.0
depyf 0.20.0
diffusers 0.36.0
dill 0.4.0
diskcache 5.6.3
distro 1.9.0
dnspython 2.8.0
docstring-parser 0.17.0
einops 0.8.2
email-validator 2.3.0
executing 2.2.1
fastapi 0.133.0
fastapi-cli 0.0.24
fastapi-cloud-cli 0.13.0
fastar 0.8.0
filelock 3.24.3
flashinfer-cubin 0.6.3
flashinfer-python 0.6.3
frozenlist 1.8.0
fsspec 2025.10.0
gguf 0.17.1
googleapis-common-protos 1.72.0
grpcio 1.78.1
grpcio-health-checking 1.78.1
grpcio-reflection 1.78.1
h11 0.16.0
hf-transfer 0.1.9
hf-xet 1.3.1
httpcore 1.0.9
httptools 0.7.1
httpx 0.28.1
httpx-sse 0.4.3
huggingface-hub 1.4.1
humanize 4.15.0
idna 3.11
ijson 3.5.0
imageio 2.36.0
imageio-ffmpeg 0.5.1
importlib-metadata 8.7.1
interegular 0.3.3
ipython 9.10.0
ipython-pygments-lexers 1.1.1
jedi 0.19.2
jinja2 3.1.6
jiter 0.13.0
jmespath 1.1.0
jsonschema 4.26.0
jsonschema-specifications 2025.9.1
lark 1.2.2
llguidance 0.7.30
llvmlite 0.44.0
lm-format-enforcer 0.11.3
loguru 0.7.3
lxml 6.0.2
markdown-it-py 4.0.0
markupsafe 3.0.3
matplotlib-inline 0.2.1
mcp 1.26.0
mdurl 0.1.2
mistral-common 1.9.1
model-hosting-container-standards 0.1.13
modelscope 1.34.0
moviepy 2.2.1
mpmath 1.3.0
msgpack 1.1.2
msgspec 0.20.0
multidict 6.7.1
multiprocess 0.70.18
nest-asyncio 1.6.0
networkx 3.6.1
ninja 1.13.0
numba 0.61.2
numpy 2.4.2
nvidia-cublas-cu12 12.8.4.1
nvidia-cuda-cupti-cu12 12.8.90
nvidia-cuda-nvrtc-cu12 12.8.93
nvidia-cuda-runtime-cu12 12.8.90
nvidia-cudnn-cu12 9.16.0.29
nvidia-cudnn-frontend 1.18.0
nvidia-cufft-cu12 11.3.3.83
nvidia-cufile-cu12 1.13.1.3
nvidia-curand-cu12 10.3.9.90
nvidia-cusolver-cu12 11.7.3.90
nvidia-cusparse-cu12 12.5.8.93
nvidia-cusparselt-cu12 0.7.1
nvidia-cutlass-dsl 4.3.5
nvidia-cutlass-dsl-libs-base 4.4.0
nvidia-ml-py 13.590.48
nvidia-nccl-cu12 2.27.5
nvidia-nvjitlink-cu12 12.8.93
nvidia-nvshmem-cu12 3.3.20
nvidia-nvtx-cu12 12.8.90
openai 2.6.1
openai-harmony 0.0.4
opencv-python-headless 4.10.0.84
opentelemetry-api 1.39.1
opentelemetry-exporter-otlp 1.39.1
opentelemetry-exporter-otlp-proto-common 1.39.1
opentelemetry-exporter-otlp-proto-grpc 1.39.1
opentelemetry-exporter-otlp-proto-http 1.39.1
opentelemetry-proto 1.39.1
opentelemetry-sdk 1.39.1
opentelemetry-semantic-conventions 0.60b1
orjson 3.11.7
outlines 0.1.11
outlines-core 0.1.26
packaging 26.0
pandas 3.0.1
parso 0.8.6
partial-json-parser 0.2.1.1.post7
pexpect 4.9.0
pillow 11.3.0
proglog 0.1.12
prometheus-client 0.24.1
prometheus-fastapi-instrumentator 7.1.0
prompt-toolkit 3.0.52
propcache 0.4.1
protobuf 6.33.5
psutil 7.2.2
ptyprocess 0.7.0
pure-eval 0.2.3
py-cpuinfo 9.0.0
py-spy 0.4.1
pyarrow 23.0.1
pybase64 1.4.3
pycountry 26.2.16
pycparser 3.0
pycryptodomex 3.23.0
pydantic 2.12.5
pydantic-core 2.41.5
pydantic-extra-types 2.11.0
pydantic-settings 2.13.1
pygments 2.19.2
pyjwt 2.11.0
pyproject-hooks 1.2.0
python-dateutil 2.9.0.post0
python-dotenv 1.2.1
python-json-logger 4.0.0
python-multipart 0.0.22
pyyaml 6.0.1
pyzmq 27.1.0
quack-kernels 0.2.4
ray 2.54.0
referencing 0.37.0
regex 2026.2.19
remote-pdb 2.1.0
requests 2.32.5
rich 14.3.3
rich-toolkit 0.19.7
rignore 0.7.6
rpds-py 0.30.0
runai-model-streamer 0.15.6
safetensors 0.7.0
scipy 1.17.1
sentencepiece 0.2.1
sentry-sdk 2.53.0
setproctitle 1.3.7
setuptools 80.10.2
sgl-kernel 0.3.21
sglang 0.5.6.post2
shellingham 1.5.4
six 1.17.0
smg-grpc-proto 0.4.0
sniffio 1.3.1
soundfile 0.13.1
sse-starlette 3.2.0
st-attn 0.0.7
stack-data 0.6.3
starlette 0.52.1
supervisor 4.3.0
sympy 1.14.0
tabulate 0.9.0
tiktoken 0.12.0
timm 1.0.16
tokenizers 0.22.2
torch 2.9.1
torch-c-dlpack-ext 0.1.5
torch-memory-saver 0.0.9
torchao 0.9.0
torchaudio 2.9.1
torchcodec 0.8.0
torchvision 0.24.1
tqdm 4.67.3
traitlets 5.14.3
transformers 5.3.0.dev0
triton 3.5.1
typer 0.24.1
typer-slim 0.24.0
typing-extensions 4.15.0
typing-inspection 0.4.2
urllib3 2.6.3
uvicorn 0.41.0
uvloop 0.22.1
vllm 0.15.1
vsa 0.0.4
watchfiles 1.1.1
wcwidth 0.6.0
websockets 16.0
xgrammar 0.1.27
xxhash 3.6.0
yarl 1.22.0
zipp 3.23.0遇到的异常: (APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325]
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325] █ █ █▄ ▄█
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325] ▄▄ ▄█ █ █ █ ▀▄▀ █ version 0.15.1
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325] █▄█▀ █ █ █ █ model Qwen/Qwen3.5-27B
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325] ▀▀ ▀▀▀▀▀ ▀▀▀▀▀ ▀ ▀
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:325]
(APIServer pid=1387856) INFO 02-25 17:16:16 [utils.py:261] non-default args: {'model_tag': 'Qwen/Qwen3.5-27B', 'api_server_count': 1, 'port': 8040, 'enable_auto_tool_choice': True, 'tool_call_parser': 'qwen3_coder', 'model': 'Qwen/Qwen3.5-27B', 'max_model_len': 262144, 'reasoning_parser': 'qwen3', 'tensor_parallel_size': 4}
(APIServer pid=1387856) Downloading Model from https://www.modelscope.cn to directory: /data2/xli/models/Qwen/Qwen3.5-27B
(APIServer pid=1387856) 2026-02-25 17:16:17,184 - modelscope - INFO - Target directory already exists, skipping creation.
(APIServer pid=1387856) Downloading Model from https://www.modelscope.cn to directory: /data2/xli/models/Qwen/Qwen3.5-27B
(APIServer pid=1387856) 2026-02-25 17:16:18,332 - modelscope - INFO - Target directory already exists, skipping creation.
(APIServer pid=1387856) Downloading Model from https://www.modelscope.cn to directory: /data2/xli/models/Qwen/Qwen3.5-27B
(APIServer pid=1387856) 2026-02-25 17:16:19,250 - modelscope - INFO - Target directory already exists, skipping creation.
(APIServer pid=1387856) Unrecognized keys in `rope_parameters` for 'rope_type'='default': {'mrope_section', 'mrope_interleaved'}
(APIServer pid=1387856) Unrecognized keys in `rope_parameters` for 'rope_type'='default': {'mrope_section', 'mrope_interleaved'}
(APIServer pid=1387856) Traceback (most recent call last):
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/bin/vllm", line 10, in <module>
(APIServer pid=1387856) sys.exit(main())
(APIServer pid=1387856) ^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/cli/main.py", line 73, in main
(APIServer pid=1387856) args.dispatch_function(args)
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/cli/serve.py", line 111, in cmd
(APIServer pid=1387856) uvloop.run(run_server(args))
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 96, in run
(APIServer pid=1387856) return __asyncio.run(
(APIServer pid=1387856) ^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/.local/share/uv/python/cpython-3.12.11-linux-x86_64-gnu/lib/python3.12/asyncio/runners.py", line 195, in run
(APIServer pid=1387856) return runner.run(main)
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/.local/share/uv/python/cpython-3.12.11-linux-x86_64-gnu/lib/python3.12/asyncio/runners.py", line 118, in run
(APIServer pid=1387856) return self._loop.run_until_complete(task)
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "uvloop/loop.pyx", line 1518, in uvloop.loop.Loop.run_until_complete
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/uvloop/__init__.py", line 48, in wrapper
(APIServer pid=1387856) return await main
(APIServer pid=1387856) ^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 919, in run_server
(APIServer pid=1387856) await run_server_worker(listen_address, sock, args, **uvicorn_kwargs)
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 938, in run_server_worker
(APIServer pid=1387856) async with build_async_engine_client(
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/.local/share/uv/python/cpython-3.12.11-linux-x86_64-gnu/lib/python3.12/contextlib.py", line 210, in __aenter__
(APIServer pid=1387856) return await anext(self.gen)
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 147, in build_async_engine_client
(APIServer pid=1387856) async with build_async_engine_client_from_engine_args(
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/.local/share/uv/python/cpython-3.12.11-linux-x86_64-gnu/lib/python3.12/contextlib.py", line 210, in __aenter__
(APIServer pid=1387856) return await anext(self.gen)
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/entrypoints/openai/api_server.py", line 173, in build_async_engine_client_from_engine_args
(APIServer pid=1387856) vllm_config = engine_args.create_engine_config(usage_context=usage_context)
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/engine/arg_utils.py", line 1374, in create_engine_config
(APIServer pid=1387856) model_config = self.create_model_config()
(APIServer pid=1387856) ^^^^^^^^^^^^^^^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/vllm/engine/arg_utils.py", line 1228, in create_model_config
(APIServer pid=1387856) return ModelConfig(
(APIServer pid=1387856) ^^^^^^^^^^^^
(APIServer pid=1387856) File "/data1/users/xli/lee_A800/Qwen-3_5/.venv/lib/python3.12/site-packages/pydantic/_internal/_dataclasses.py", line 121, in __init__
(APIServer pid=1387856) s.__pydantic_validator__.validate_python(ArgsKwargs(args, kwargs), self_instance=s)
(APIServer pid=1387856) pydantic_core._pydantic_core.ValidationError: 1 validation error for ModelConfig
(APIServer pid=1387856) Value error, Model architectures ['Qwen3_5ForConditionalGeneration'] are not supported for now. Supported architectures: dict_keys(['AfmoeForCausalLM', 'ApertusForCausalLM', 'AquilaModel', 'AquilaForCausalLM', 'ArceeForCausalLM', 'ArcticForCausalLM', 'BaiChuanForCausalLM', 'BaichuanForCausalLM', 'BailingMoeForCausalLM', 'BailingMoeV2ForCausalLM', 'BambaForCausalLM', 'BloomForCausalLM', 'ChatGLMModel', 'ChatGLMForConditionalGeneration', 'CohereForCausalLM', 'Cohere2ForCausalLM', 'CwmForCausalLM', 'DbrxForCausalLM', 'DeciLMForCausalLM', 'DeepseekForCausalLM', 'DeepseekV2ForCausalLM', 'DeepseekV3ForCausalLM', 'DeepseekV32ForCausalLM', 'Dots1ForCausalLM', 'Ernie4_5ForCausalLM', 'Ernie4_5_MoeForCausalLM', 'ExaoneForCausalLM', 'Exaone4ForCausalLM', 'ExaoneMoEForCausalLM', 'Fairseq2LlamaForCausalLM', 'FalconForCausalLM', 'FalconMambaForCausalLM', 'FalconH1ForCausalLM', 'FlexOlmoForCausalLM', 'GemmaForCausalLM', 'Gemma2ForCausalLM', 'Gemma3ForCausalLM', 'Gemma3nForCausalLM', 'Qwen3NextForCausalLM', 'GlmForCausalLM', 'Glm4ForCausalLM', 'Glm4MoeForCausalLM', 'Glm4MoeLiteForCausalLM', 'GptOssForCausalLM', 'GPT2LMHeadModel', 'GPTBigCodeForCausalLM', 'GPTJForCausalLM', 'GPTNeoXForCausalLM', 'GraniteForCausalLM', 'GraniteMoeForCausalLM', 'GraniteMoeHybridForCausalLM', 'GraniteMoeSharedForCausalLM', 'GritLM', 'Grok1ModelForCausalLM', 'Grok1ForCausalLM', 'HunYuanMoEV1ForCausalLM', 'HunYuanDenseV1ForCausalLM', 'HCXVisionForCausalLM', 'InternLMForCausalLM', 'InternLM2ForCausalLM', 'InternLM2VEForCausalLM', 'InternLM3ForCausalLM', 'IQuestCoderForCausalLM', 'IQuestLoopCoderForCausalLM', 'JAISLMHeadModel', 'Jais2ForCausalLM', 'JambaForCausalLM', 'KimiLinearForCausalLM', 'Lfm2ForCausalLM', 'Lfm2MoeForCausalLM', 'LlamaForCausalLM', 'Llama4ForCausalLM', 'LLaMAForCausalLM', 'LongcatFlashForCausalLM', 'MambaForCausalLM', 'Mamba2ForCausalLM', 'MiniCPMForCausalLM', 'MiniCPM3ForCausalLM', 'MiniMaxForCausalLM', 'MiniMaxText01ForCausalLM', 'MiniMaxM1ForCausalLM', 'MiniMaxM2ForCausalLM', 'MistralForCausalLM', 'MistralLarge3ForCausalLM', 'MixtralForCausalLM', 'MptForCausalLM', 'MPTForCausalLM', 'MiMoForCausalLM', 'MiMoV2FlashForCausalLM', 'NemotronForCausalLM', 'NemotronHForCausalLM', 'OlmoForCausalLM', 'Olmo2ForCausalLM', 'Olmo3ForCausalLM', 'OlmoeForCausalLM', 'OPTForCausalLM', 'OrionForCausalLM', 'OuroForCausalLM', 'PanguEmbeddedForCausalLM', 'PanguProMoEV2ForCausalLM', 'PanguUltraMoEForCausalLM', 'PersimmonForCausalLM', 'PhiForCausalLM', 'Phi3ForCausalLM', 'PhiMoEForCausalLM', 'Plamo2ForCausalLM', 'Plamo3ForCausalLM', 'QWenLMHeadModel', 'Qwen2ForCausalLM', 'Qwen2MoeForCausalLM', 'Qwen3ForCausalLM', 'Qwen3MoeForCausalLM', 'RWForCausalLM', 'SeedOssForCausalLM', 'Step1ForCausalLM', 'Step3TextForCausalLM', 'Step3p5ForCausalLM', 'StableLMEpochForCausalLM', 'StableLmForCausalLM', 'Starcoder2ForCausalLM', 'SolarForCausalLM', 'TeleChatForCausalLM', 'TeleChat2ForCausalLM', 'TeleFLMForCausalLM', 'XverseForCausalLM', 'Zamba2ForCausalLM', 'BertModel', 'BertSpladeSparseEmbeddingModel', 'Gemma2Model', 'Gemma3TextModel', 'GPT2ForSequenceClassification', 'GteModel', 'GteNewModel', 'InternLM2ForRewardModel', 'JambaForSequenceClassification', 'LlamaBidirectionalModel', 'LlamaModel', 'MistralModel', 'ModernBertModel', 'NomicBertModel', 'Qwen2Model', 'Qwen2ForRewardModel', 'Qwen2ForProcessRewardModel', 'RobertaForMaskedLM', 'RobertaModel', 'XLMRobertaModel', 'BgeM3EmbeddingModel', 'CLIPModel', 'LlavaNextForConditionalGeneration', 'Phi3VForCausalLM', 'Qwen2VLForConditionalGeneration', 'SiglipModel', 'PrithviGeoSpatialMAE', 'Terratorch', 'BertForSequenceClassification', 'BertForTokenClassification', 'GteNewForSequenceClassification', 'JinaVLForRanking', 'LlamaBidirectionalForSequenceClassification', 'ModernBertForSequenceClassification', 'ModernBertForTokenClassification', 'RobertaForSequenceClassification', 'XLMRobertaForSequenceClassification', 'AriaForConditionalGeneration', 'AudioFlamingo3ForConditionalGeneration', 'AyaVisionForConditionalGeneration', 'BagelForConditionalGeneration', 'BeeForConditionalGeneration', 'Blip2ForConditionalGeneration', 'ChameleonForConditionalGeneration', 'Cohere2VisionForConditionalGeneration', 'DeepseekVLV2ForCausalLM', 'DeepseekOCRForCausalLM', 'DotsOCRForCausalLM', 'Eagle2_5_VLForConditionalGeneration', 'Ernie4_5_VLMoeForConditionalGeneration', 'FuyuForCausalLM', 'Gemma3ForConditionalGeneration', 'Gemma3nForConditionalGeneration', 'GlmAsrForConditionalGeneration', 'GLM4VForCausalLM', 'Glm4vForConditionalGeneration', 'Glm4vMoeForConditionalGeneration', 'GraniteSpeechForConditionalGeneration', 'H2OVLChatModel', 'HunYuanVLForConditionalGeneration', 'StepVLForConditionalGeneration', 'InternVLChatModel', 'NemotronH_Nano_VL_V2', 'OpenCUAForConditionalGeneration', 'InternS1ForConditionalGeneration', 'InternVLForConditionalGeneration', 'Idefics3ForConditionalGeneration', 'IsaacForConditionalGeneration', 'SmolVLMForConditionalGeneration', 'KananaVForConditionalGeneration', 'KeyeForConditionalGeneration', 'KeyeVL1_5ForConditionalGeneration', 'RForConditionalGeneration', 'KimiVLForConditionalGeneration', 'KimiK25ForConditionalGeneration', 'LightOnOCRForConditionalGeneration', 'Lfm2VlForConditionalGeneration', 'Llama_Nemotron_Nano_VL', 'Llama4ForConditionalGeneration', 'LlavaForConditionalGeneration', 'LlavaNextVideoForConditionalGeneration', 'LlavaOnevisionForConditionalGeneration', 'MantisForConditionalGeneration', 'MiDashengLMModel', 'MiniMaxVL01ForConditionalGeneration', 'MiniCPMO', 'MiniCPMV', 'Mistral3ForConditionalGeneration', 'MolmoForCausalLM', 'Molmo2ForConditionalGeneration', 'NVLM_D', 'Ovis', 'Ovis2_5', 'PaddleOCRVLForConditionalGeneration', 'PaliGemmaForConditionalGeneration', 'Phi4MMForCausalLM', 'PixtralForConditionalGeneration', 'QwenVLForConditionalGeneration', 'Qwen2_5_VLForConditionalGeneration', 'Qwen2AudioForConditionalGeneration', 'Qwen2_5OmniModel', 'Qwen2_5OmniForConditionalGeneration', 'Qwen3OmniMoeForConditionalGeneration', 'Qwen3VLForConditionalGeneration', 'Qwen3VLMoeForConditionalGeneration', 'SkyworkR1VChatModel', 'Step3VLForConditionalGeneration', 'TarsierForConditionalGeneration', 'Tarsier2ForConditionalGeneration', 'UltravoxModel', 'VoxtralForConditionalGeneration', 'VoxtralStreamingGeneration', 'NemotronParseForConditionalGeneration', 'WhisperForConditionalGeneration', 'MiMoMTPModel', 'EagleLlamaForCausalLM', 'EagleLlama4ForCausalLM', 'EagleMiniCPMForCausalLM', 'Eagle3LlamaForCausalLM', 'LlamaForCausalLMEagle3', 'Eagle3Qwen2_5vlForCausalLM', 'Eagle3Qwen3vlForCausalLM', 'EagleMistralLarge3ForCausalLM', 'EagleDeepSeekMTPModel', 'DeepSeekMTPModel', 'ErnieMTPModel', 'ExaoneMoeMTP', 'LongCatFlashMTPModel', 'Glm4MoeMTPModel', 'Glm4MoeLiteMTPModel', 'MedusaModel', 'OpenPanguMTPModel', 'Qwen3NextMTP', 'Step3p5MTP', 'SmolLM3ForCausalLM', 'Emu3ForConditionalGeneration', 'TransformersForCausalLM', 'TransformersMoEForCausalLM', 'TransformersMultiModalForCausalLM', 'TransformersMultiModalMoEForCausalLM', 'TransformersEmbeddingModel', 'TransformersMoEEmbeddingModel', 'TransformersMultiModalEmbeddingModel', 'TransformersForSequenceClassification', 'TransformersMoEForSequenceClassification', 'TransformersMultiModalForSequenceClassification']) [type=value_error, input_value=ArgsKwargs((), {'model': ...rocessor_plugin': None}), input_type=ArgsKwargs]
(APIServer pid=1387856) For further information visit https://errors.pydantic.dev/2.12/v/value_error有无大佬指点一下是不是哪个包不合适? |
Beta Was this translation helpful? Give feedback.
Answered by
JJJYmmm
Mar 5, 2026
Replies: 4 comments 2 replies
-
|
支持qwen3.5的vllm还没放release包,目前需要clone下来vllm的main分支,然后install from src,或者等vllm他们肝出新release |
Beta Was this translation helpful? Give feedback.
1 reply
-
|
也可以直接用nightly镜像,或者等vllm 0.17.0 |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
JJJYmmm
-
|
这个问题解决了吗 |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
我用的VLLM=0.17.1,还是不行,报错类似,有大佬指教一下吗 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
也可以直接用nightly镜像,或者等vllm 0.17.0