|
1 | 1 | huggingface_mini_db = { |
2 | | - "starcoder/15b/base": { |
3 | | - "backend": "autogptq", |
4 | | - "model_path": "TheBloke/starcoder-GPTQ", |
5 | | - "model_class_kwargs": {}, |
6 | | - "required_memory_mb": 18000, |
7 | | - "T": 4096, |
8 | | - "filter_caps": ["completion"], |
9 | | - "deprecated": True, |
10 | | - }, |
11 | | - "starcoder/15b/plus": { |
12 | | - "backend": "autogptq", |
13 | | - "model_path": "TheBloke/starcoderplus-GPTQ", |
14 | | - "model_class_kwargs": {}, |
15 | | - "required_memory_mb": 18000, |
16 | | - "T": 4096, |
17 | | - "filter_caps": ["completion"], |
18 | | - "deprecated": True, |
19 | | - }, |
20 | | - "starchat/15b/beta": { |
21 | | - "backend": "autogptq", |
22 | | - "model_path": "TheBloke/starchat-beta-GPTQ", |
23 | | - "model_class_kwargs": {}, |
24 | | - "required_memory_mb": 18000, |
25 | | - "T": 4096, |
26 | | - "filter_caps": ["chat"], |
27 | | - "deprecated": True, |
28 | | - }, |
29 | | - "starcoder/1b/base": { |
30 | | - "backend": "transformers", |
31 | | - "model_path": "smallcloudai/starcoderbase-1b", |
32 | | - "model_class_kwargs": {}, |
33 | | - "required_memory_mb": 8000, |
34 | | - "T": 8192, |
35 | | - "filter_caps": ["completion", "finetune"], |
36 | | - "deprecated": True, |
37 | | - }, |
38 | | - "starcoder/3b/base": { |
39 | | - "backend": "transformers", |
40 | | - "model_path": "smallcloudai/starcoderbase-3b", |
41 | | - "model_class_kwargs": {}, |
42 | | - "required_memory_mb": 12000, |
43 | | - "T": 4096, |
44 | | - "filter_caps": ["completion", "finetune"], |
45 | | - "deprecated": True, |
46 | | - }, |
47 | | - "starcoder/7b/base": { |
48 | | - "backend": "transformers", |
49 | | - "model_path": "smallcloudai/starcoderbase-7b", |
50 | | - "model_class_kwargs": {}, |
51 | | - "required_memory_mb": 20000, |
52 | | - "T": 4096, |
53 | | - "filter_caps": ["completion", "finetune"], |
54 | | - "deprecated": True, |
55 | | - }, |
56 | | - "wizardcoder/15b": { |
57 | | - "backend": "autogptq", |
58 | | - "model_path": "TheBloke/WizardCoder-15B-1.0-GPTQ", |
59 | | - "model_class_kwargs": {}, |
60 | | - "required_memory_mb": 18000, |
61 | | - "T": 4096, |
62 | | - "filter_caps": ["completion"], |
63 | | - "deprecated": True, |
64 | | - }, |
65 | | - "wizardlm/7b": { |
66 | | - "backend": "autogptq", |
67 | | - "model_path": "TheBloke/WizardLM-7B-V1.0-Uncensored-GPTQ", |
68 | | - "model_class_kwargs": {}, |
69 | | - "required_memory_mb": 8000, |
70 | | - "T": 2048, |
71 | | - "filter_caps": ["chat"], |
72 | | - "deprecated": True, |
73 | | - }, |
74 | | - "wizardlm/13b": { |
75 | | - "backend": "autogptq", |
76 | | - "model_path": "TheBloke/WizardLM-13B-V1.1-GPTQ", |
77 | | - "model_class_kwargs": {}, |
78 | | - "required_memory_mb": 14000, |
79 | | - "T": 2048, |
80 | | - "filter_caps": ["chat"], |
81 | | - "deprecated": True, |
82 | | - }, |
83 | | - "llama2/7b": { |
84 | | - "backend": "autogptq", |
85 | | - "model_path": "TheBloke/Llama-2-7b-Chat-GPTQ", |
86 | | - "model_class_kwargs": {}, |
87 | | - "required_memory_mb": 8000, |
88 | | - "T": 2048, |
89 | | - "filter_caps": ["chat"], |
90 | | - "deprecated": True, |
91 | | - }, |
92 | | - "llama2/13b": { |
93 | | - "backend": "autogptq", |
94 | | - "model_path": "TheBloke/Llama-2-13B-chat-GPTQ", |
95 | | - "model_class_kwargs": {}, |
96 | | - "required_memory_mb": 14000, |
97 | | - "T": 2048, |
98 | | - "filter_caps": ["chat"], |
99 | | - "deprecated": True, |
100 | | - }, |
101 | | - "codellama/7b": { |
102 | | - "backend": "transformers", |
103 | | - "model_path": "TheBloke/CodeLlama-7B-fp16", |
104 | | - "model_class_kwargs": {}, |
105 | | - "required_memory_mb": 14000, |
106 | | - "T": 2048, |
107 | | - "filter_caps": ["completion"], |
108 | | - "deprecated": True, |
109 | | - }, |
110 | | - "wizardlm/30b": { |
111 | | - "backend": "transformers", |
112 | | - "model_path": "TheBloke/WizardLM-30B-fp16", |
113 | | - "model_class_kwargs": { |
114 | | - "load_in_4bit": True, |
115 | | - }, |
116 | | - "T": 2048, |
117 | | - "filter_caps": ["chat"], |
118 | | - "deprecated": True, |
119 | | - }, |
120 | 2 | "deepseek-coder/1.3b/base": { |
121 | 3 | "backend": "transformers", |
122 | 4 | "model_path": "deepseek-ai/deepseek-coder-1.3b-base", |
|
0 commit comments