Skip to content

Commit 0c05963

Browse files
authored
[Model] Deliver Qwen2.5-Coder 0.5B and 3B (#627)
This PR delivers Qwen2.5-Coder 0.5B and 3B, as 1.5B and 7B are already supported. We add the following models to the model list: - `Qwen2.5-Coder-3B-Instruct-q4f16_1-MLC` - `Qwen2.5-Coder-3B-Instruct-q4f32_1-MLC` - `Qwen2.5-Coder-0.5B-Instruct-q4f16_1-MLC` - `Qwen2.5-Coder-0.5B-Instruct-q4f32_1-MLC` - `Qwen2.5-Coder-0.5B-Instruct-q0f16-MLC` - `Qwen2.5-Coder-0.5B-Instruct-q0f32-MLC` Note that the official `config.json` of the coder models are identical to that of the chat models, so we do not need to compile new WASMs.
1 parent 5fa741f commit 0c05963

File tree

1 file changed

+84
-0
lines changed

1 file changed

+84
-0
lines changed

src/config.ts

Lines changed: 84 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1069,6 +1069,62 @@ export const prebuiltAppConfig: AppConfig = {
10691069
},
10701070
},
10711071
// Qwen2.5-Coder
1072+
{
1073+
model:
1074+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-0.5B-Instruct-q4f16_1-MLC",
1075+
model_id: "Qwen2.5-Coder-0.5B-Instruct-q4f16_1-MLC",
1076+
model_lib:
1077+
modelLibURLPrefix +
1078+
modelVersion +
1079+
"/Qwen2-0.5B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm",
1080+
low_resource_required: true,
1081+
vram_required_MB: 944.62,
1082+
overrides: {
1083+
context_window_size: 4096,
1084+
},
1085+
},
1086+
{
1087+
model:
1088+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-0.5B-Instruct-q4f32_1-MLC",
1089+
model_id: "Qwen2.5-Coder-0.5B-Instruct-q4f32_1-MLC",
1090+
model_lib:
1091+
modelLibURLPrefix +
1092+
modelVersion +
1093+
"/Qwen2-0.5B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm",
1094+
low_resource_required: true,
1095+
vram_required_MB: 1060.2,
1096+
overrides: {
1097+
context_window_size: 4096,
1098+
},
1099+
},
1100+
{
1101+
model:
1102+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-0.5B-Instruct-q0f16-MLC",
1103+
model_id: "Qwen2.5-Coder-0.5B-Instruct-q0f16-MLC",
1104+
model_lib:
1105+
modelLibURLPrefix +
1106+
modelVersion +
1107+
"/Qwen2-0.5B-Instruct-q0f16-ctx4k_cs1k-webgpu.wasm",
1108+
low_resource_required: true,
1109+
vram_required_MB: 1624.12,
1110+
overrides: {
1111+
context_window_size: 4096,
1112+
},
1113+
},
1114+
{
1115+
model:
1116+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-0.5B-Instruct-q0f32-MLC",
1117+
model_id: "Qwen2.5-Coder-0.5B-Instruct-q0f32-MLC",
1118+
model_lib:
1119+
modelLibURLPrefix +
1120+
modelVersion +
1121+
"/Qwen2-0.5B-Instruct-q0f32-ctx4k_cs1k-webgpu.wasm",
1122+
low_resource_required: true,
1123+
vram_required_MB: 2654.75,
1124+
overrides: {
1125+
context_window_size: 4096,
1126+
},
1127+
},
10721128
{
10731129
model:
10741130
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-1.5B-Instruct-q4f16_1-MLC",
@@ -1097,6 +1153,34 @@ export const prebuiltAppConfig: AppConfig = {
10971153
context_window_size: 4096,
10981154
},
10991155
},
1156+
{
1157+
model:
1158+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-3B-Instruct-q4f16_1-MLC",
1159+
model_id: "Qwen2.5-Coder-3B-Instruct-q4f16_1-MLC",
1160+
model_lib:
1161+
modelLibURLPrefix +
1162+
modelVersion +
1163+
"/Qwen2.5-3B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm",
1164+
low_resource_required: true,
1165+
vram_required_MB: 2504.76,
1166+
overrides: {
1167+
context_window_size: 4096,
1168+
},
1169+
},
1170+
{
1171+
model:
1172+
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-3B-Instruct-q4f32_1-MLC",
1173+
model_id: "Qwen2.5-Coder-3B-Instruct-q4f32_1-MLC",
1174+
model_lib:
1175+
modelLibURLPrefix +
1176+
modelVersion +
1177+
"/Qwen2.5-3B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm",
1178+
low_resource_required: true,
1179+
vram_required_MB: 2893.64,
1180+
overrides: {
1181+
context_window_size: 4096,
1182+
},
1183+
},
11001184
{
11011185
model:
11021186
"https://huggingface.co/mlc-ai/Qwen2.5-Coder-7B-Instruct-q4f16_1-MLC",

0 commit comments

Comments
 (0)