Skip to content

Commit 415d109

Browse files
authored
[Fix] Update Supported Models List (#1690)
1 parent 521b35f commit 415d109

File tree

2 files changed

+11
-0
lines changed

2 files changed

+11
-0
lines changed

README.md

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -49,6 +49,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi
4949
- Aquila & Aquila2 (`BAAI/AquilaChat2-7B`, `BAAI/AquilaChat2-34B`, `BAAI/Aquila-7B`, `BAAI/AquilaChat-7B`, etc.)
5050
- Baichuan (`baichuan-inc/Baichuan-7B`, `baichuan-inc/Baichuan-13B-Chat`, etc.)
5151
- BLOOM (`bigscience/bloom`, `bigscience/bloomz`, etc.)
52+
- ChatGLM (`THUDM/chatglm2-6b`, `THUDM/chatglm3-6b`, etc.)
5253
- Falcon (`tiiuae/falcon-7b`, `tiiuae/falcon-40b`, `tiiuae/falcon-rw-7b`, etc.)
5354
- GPT-2 (`gpt2`, `gpt2-xl`, etc.)
5455
- GPT BigCode (`bigcode/starcoder`, `bigcode/gpt_bigcode-santacoder`, etc.)
@@ -61,6 +62,7 @@ vLLM seamlessly supports many Hugging Face models, including the following archi
6162
- OPT (`facebook/opt-66b`, `facebook/opt-iml-max-30b`, etc.)
6263
- Phi-1.5 (`microsoft/phi-1_5`, etc.)
6364
- Qwen (`Qwen/Qwen-7B`, `Qwen/Qwen-7B-Chat`, etc.)
65+
- Yi (`01-ai/Yi-6B`, `01-ai/Yi-34B`, etc.)
6466

6567
Install vLLM with pip or [from source](https://vllm.readthedocs.io/en/latest/getting_started/installation.html#build-from-source):
6668

docs/source/models/supported_models.rst

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,9 @@ Alongside each architecture, we include some popular models that use it.
2020
* - :code:`BaiChuanForCausalLM`
2121
- Baichuan
2222
- :code:`baichuan-inc/Baichuan-7B`, :code:`baichuan-inc/Baichuan-13B-Chat`, etc.
23+
* - :code:`ChatGLMModel`
24+
- ChatGLM
25+
- :code:`THUDM/chatglm2-6b`, :code:`THUDM/chatglm3-6b`, etc.
2326
* - :code:`BloomForCausalLM`
2427
- BLOOM, BLOOMZ, BLOOMChat
2528
- :code:`bigscience/bloom`, :code:`bigscience/bloomz`, etc.
@@ -53,9 +56,15 @@ Alongside each architecture, we include some popular models that use it.
5356
* - :code:`OPTForCausalLM`
5457
- OPT, OPT-IML
5558
- :code:`facebook/opt-66b`, :code:`facebook/opt-iml-max-30b`, etc.
59+
* - :code:`PhiForCausalLM`
60+
- Phi-1.5
61+
- :code:`microsoft/phi-1_5`, etc.
5662
* - :code:`QWenLMHeadModel`
5763
- Qwen
5864
- :code:`Qwen/Qwen-7B`, :code:`Qwen/Qwen-7B-Chat`, etc.
65+
* - :code:`YiForCausalLM`
66+
- Yi
67+
- :code:`01-ai/Yi-6B`, :code:`01-ai/Yi-34B`, etc.
5968

6069
If your model uses one of the above model architectures, you can seamlessly run your model with vLLM.
6170
Otherwise, please refer to :ref:`Adding a New Model <adding_a_new_model>` for instructions on how to implement support for your model.

0 commit comments

Comments
 (0)