You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
|`DbrxForCausalLM`| DBRX |`databricks/dbrx-base`, `databricks/dbrx-instruct`, etc. || ✅︎ | ✅︎ |
@@ -426,9 +424,6 @@ Some models are supported only via the [Transformers backend](#transformers). Th
426
424
!!! note
427
425
Currently, the ROCm version of vLLM supports Mistral and Mixtral only for context lengths up to 4096.
428
426
429
-
!!! note
430
-
Some mBART models' config files do not have an `architecture` defined. Therefore, you need to use `--hf-overrides '{"architectures": ["MBartForConditionalGeneration"]}'` to explicitly specify the use of the `MBartForConditionalGeneration` architecture.
431
-
432
427
### Pooling Models
433
428
434
429
See [this page](./pooling_models.md) for more information on how to use pooling models.
@@ -625,9 +620,7 @@ These models primarily accept the [`LLM.generate`](./generative_models.md#llmgen
625
620
|`ChameleonForConditionalGeneration`| Chameleon | T + I |`facebook/chameleon-7b`, etc. || ✅︎ | ✅︎ |
626
621
|`Cohere2VisionForConditionalGeneration`| Command A Vision | T + I<sup>+</sup> |`CohereLabs/command-a-vision-07-2025`, etc. || ✅︎ | ✅︎ |
627
622
|`DeepseekVLV2ForCausalLM`<sup>^</sup> | DeepSeek-VL2 | T + I<sup>+</sup> |`deepseek-ai/deepseek-vl2-tiny`, `deepseek-ai/deepseek-vl2-small`, `deepseek-ai/deepseek-vl2`, etc. || ✅︎ | ✅︎ |
628
-
|`DonutForConditionalGeneration`<sup>^</sup> | Donut | T + I |`ByteDance/Dolphin`, `naver-clova-ix/donut-base-finetuned-docvqa`, etc. ||||
0 commit comments