You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/models/supported_models.md
+25-3Lines changed: 25 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -626,7 +626,29 @@ See [this page](../features/multimodal_inputs.md) on how to pass multi-modal inp
626
626
For hybrid-only models such as Llama-4, Step3 and Mistral-3, a text-only mode can be enabled by setting all supported multimodal modalities to 0 (e.g, `--limit-mm-per-prompt '{"image":0}`) so that their multimodal modules will not be loaded to free up more GPU memory for KV cache.
627
627
628
628
!!! note
629
-
vLLM currently only supports adding LoRA to the language backbone of multimodal models.
629
+
vLLM currently only supports dynamic LoRA adapters on the language backbone of multimodal models.
630
+
If you wish to use a model with LoRA in the multi-modal encoder,
631
+
please merge the weights into the base model first before running it in vLLM like a regular model.
632
+
633
+
```python
634
+
from peft import PeftConfig, PeftModel
635
+
from transformers import AutoModelForImageTextToText, AutoProcessor
0 commit comments