@@ -314,6 +314,8 @@ Specified using `--task generate`.
314
314
| ` DeepseekForCausalLM ` | DeepSeek | ` deepseek-ai/deepseek-llm-67b-base ` , ` deepseek-ai/deepseek-llm-7b-chat ` etc. | | ✅︎ |
315
315
| ` DeepseekV2ForCausalLM ` | DeepSeek-V2 | ` deepseek-ai/DeepSeek-V2 ` , ` deepseek-ai/DeepSeek-V2-Chat ` etc. | | ✅︎ |
316
316
| ` DeepseekV3ForCausalLM ` | DeepSeek-V3 | ` deepseek-ai/DeepSeek-V3-Base ` , ` deepseek-ai/DeepSeek-V3 ` etc. | | ✅︎ |
317
+ | ` Ernie4_5ForCausalLM ` | Ernie4.5 | ` baidu/ERNIE-4.5-0.3B-PT ` ,etc. | | ✅︎ | ✅︎ |
318
+ | ` Ernie4_5_MoeForCausalLM ` | Ernie4.5MoE | ` baidu/ERNIE-4.5-21B-A3B-PT ` , ` baidu/ERNIE-4.5-300B-A47B-PT ` , etc. | | ✅︎ | ✅︎ |
317
319
| ` ExaoneForCausalLM ` | EXAONE-3 | ` LGAI-EXAONE/EXAONE-3.0-7.8B-Instruct ` , etc. | ✅︎ | ✅︎ |
318
320
| ` FalconForCausalLM ` | Falcon | ` tiiuae/falcon-7b ` , ` tiiuae/falcon-40b ` , ` tiiuae/falcon-rw-7b ` , etc. | | ✅︎ |
319
321
| ` FalconMambaForCausalLM ` | FalconMamba | ` tiiuae/falcon-mamba-7b ` , ` tiiuae/falcon-mamba-7b-instruct ` , etc. | | ✅︎ |
@@ -371,7 +373,6 @@ Specified using `--task generate`.
371
373
| ` XverseForCausalLM ` | XVERSE | ` xverse/XVERSE-7B-Chat ` , ` xverse/XVERSE-13B-Chat ` , ` xverse/XVERSE-65B-Chat ` , etc. | ✅︎ | ✅︎ |
372
374
| ` MiniMaxText01ForCausalLM ` | MiniMax-Text | ` MiniMaxAI/MiniMax-Text-01 ` , etc. | | |
373
375
| ` Zamba2ForCausalLM ` | Zamba2 | ` Zyphra/Zamba2-7B-instruct ` , ` Zyphra/Zamba2-2.7B-instruct ` , ` Zyphra/Zamba2-1.2B-instruct ` , etc. | | |
374
-
375
376
!!! note
376
377
Currently, the ROCm version of vLLM supports Mistral and Mixtral only for context lengths up to 4096.
377
378
@@ -556,10 +557,10 @@ Specified using `--task generate`.
556
557
| ` SmolVLMForConditionalGeneration ` | SmolVLM2 | T + I | ` SmolVLM2-2.2B-Instruct ` | ✅︎ | | ✅︎ |
557
558
| ` TarsierForConditionalGeneration ` | Tarsier | T + I<sup >E+</sup > | ` omni-search/Tarsier-7b ` ,` omni-search/Tarsier-34b ` | | ✅︎ | ✅︎ |
558
559
559
- <sup >^</sup > You need to set the architecture name via ` --hf-overrides ` to match the one in vLLM.
560
-   ;  ;  ;  ; • For example, to use DeepSeek-VL2 series models:
561
-   ;  ;  ;  ;  ;  ; ` --hf-overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}' `
562
- <sup >E</sup > Pre-computed embeddings can be inputted for this modality.
560
+ <sup >^</sup > You need to set the architecture name via ` --hf-overrides ` to match the one in vLLM.
561
+   ;  ;  ;  ; • For example, to use DeepSeek-VL2 series models:
562
+   ;  ;  ;  ;  ;  ; ` --hf-overrides '{"architectures": ["DeepseekVLV2ForCausalLM"]}' `
563
+ <sup >E</sup > Pre-computed embeddings can be inputted for this modality.
563
564
<sup >+</sup > Multiple items can be inputted per text prompt for this modality.
564
565
565
566
!!! warning
0 commit comments