@@ -20,6 +20,9 @@ Alongside each architecture, we include some popular models that use it.
20
20
* - :code: `BaiChuanForCausalLM `
21
21
- Baichuan
22
22
- :code: `baichuan-inc/Baichuan-7B `, :code: `baichuan-inc/Baichuan-13B-Chat `, etc.
23
+ * - :code: `ChatGLMModel `
24
+ - ChatGLM
25
+ - :code: `THUDM/chatglm2-6b `, :code: `THUDM/chatglm3-6b `, etc.
23
26
* - :code: `BloomForCausalLM `
24
27
- BLOOM, BLOOMZ, BLOOMChat
25
28
- :code: `bigscience/bloom `, :code: `bigscience/bloomz `, etc.
@@ -53,9 +56,15 @@ Alongside each architecture, we include some popular models that use it.
53
56
* - :code: `OPTForCausalLM `
54
57
- OPT, OPT-IML
55
58
- :code: `facebook/opt-66b `, :code: `facebook/opt-iml-max-30b `, etc.
59
+ * - :code: `PhiForCausalLM `
60
+ - Phi-1.5
61
+ - :code: `microsoft/phi-1_5 `, etc.
56
62
* - :code: `QWenLMHeadModel `
57
63
- Qwen
58
64
- :code: `Qwen/Qwen-7B `, :code: `Qwen/Qwen-7B-Chat `, etc.
65
+ * - :code: `YiForCausalLM `
66
+ - Yi
67
+ - :code: `01-ai/Yi-6B `, :code: `01-ai/Yi-34B `, etc.
59
68
60
69
If your model uses one of the above model architectures, you can seamlessly run your model with vLLM.
61
70
Otherwise, please refer to :ref: `Adding a New Model <adding_a_new_model >` for instructions on how to implement support for your model.
0 commit comments