You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/source/models/supported_models.md
+11-4Lines changed: 11 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -24,7 +24,7 @@ vLLM also supports model implementations that are available in Transformers. Thi
24
24
25
25
To check if the modeling backend is Transformers, you can simply do this:
26
26
27
-
```python
27
+
```python
28
28
from vllm importLLM
29
29
llm = LLM(model=..., task="generate") # Name or path of your model
30
30
llm.apply_model(lambdamodel: print(type(model)))
@@ -55,7 +55,7 @@ If your model is neither supported natively by vLLM or Transformers, you can sti
55
55
Simply set `trust_remote_code=True` and vLLM will run any model on the Model Hub that is compatible with Transformers.
56
56
Provided that the model writer implements their model in a compatible way, this means that you can run new models before they are officially supported in Transformers or vLLM!
57
57
58
-
```python
58
+
```python
59
59
from vllm importLLM
60
60
llm = LLM(model=..., task="generate", trust_remote_code=True) # Name or path of your model
*`THUDM/chatglm2-6b`, `THUDM/chatglm3-6b`, `ShieldLM-6B-chatglm3`, etc.
239
239
* ✅︎
240
240
* ✅︎
241
241
-*`CohereForCausalLM`, `Cohere2ForCausalLM`
@@ -850,6 +850,13 @@ See [this page](#generative-models) for more information on how to use generativ
850
850
*
851
851
* ✅︎
852
852
* ✅︎
853
+
-*`Llama4ForConditionalGeneration`
854
+
* Llama-4-17B-Omni-Instruct
855
+
* T + I<sup>+</sup>
856
+
*`meta-llama/Llama-4-Scout-17B-16E-Instruct`, `meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8`, `meta-llama/Llama-4-Maverick-17B-128E-Instruct`, etc.
0 commit comments