Skip to content

Commit db77ff5

Browse files
authored
[Llama] Add Llama3.2 Instruct 1B 3B to preset (#575)
This PR supports Llama3.2-1B and 3B Instruct. Hence we add the following Llama3.2 models to the prebuilt list: - `Llama-3.2-1B-Instruct-q4f16_1-MLC` - `Llama-3.2-1B-Instruct-q4f32_1-MLC` - `Llama-3.2-1B-Instruct-q0f16-MLC` - `Llama-3.2-1B-Instruct-q0f32-MLC` - `Llama-3.2-3B-Instruct-q4f16_1-MLC` - `Llama-3.2-3B-Instruct-q4f32_1-MLC`
1 parent 0b5f405 commit db77ff5

File tree

1 file changed

+79
-1
lines changed

1 file changed

+79
-1
lines changed

src/config.ts

Lines changed: 79 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -308,7 +308,85 @@ export const functionCallingModelIds = [
308308
export const prebuiltAppConfig: AppConfig = {
309309
useIndexedDBCache: false,
310310
model_list: [
311-
// Llama-3
311+
{
312+
model: "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q4f32_1-MLC",
313+
model_id: "Llama-3.2-1B-Instruct-q4f32_1-MLC",
314+
model_lib:
315+
modelLibURLPrefix +
316+
modelVersion +
317+
"/Llama-3.2-1B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm",
318+
vram_required_MB: 1128.82,
319+
low_resource_required: true,
320+
overrides: {
321+
context_window_size: 4096,
322+
},
323+
},
324+
{
325+
model: "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q4f16_1-MLC",
326+
model_id: "Llama-3.2-1B-Instruct-q4f16_1-MLC",
327+
model_lib:
328+
modelLibURLPrefix +
329+
modelVersion +
330+
"/Llama-3.2-1B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm",
331+
vram_required_MB: 879.04,
332+
low_resource_required: true,
333+
overrides: {
334+
context_window_size: 4096,
335+
},
336+
},
337+
{
338+
model: "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q0f32-MLC",
339+
model_id: "Llama-3.2-1B-Instruct-q0f32-MLC",
340+
model_lib:
341+
modelLibURLPrefix +
342+
modelVersion +
343+
"/Llama-3.2-1B-Instruct-q0f32-ctx4k_cs1k-webgpu.wasm",
344+
vram_required_MB: 5106.26,
345+
low_resource_required: true,
346+
overrides: {
347+
context_window_size: 4096,
348+
},
349+
},
350+
{
351+
model: "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q0f16-MLC",
352+
model_id: "Llama-3.2-1B-Instruct-q0f16-MLC",
353+
model_lib:
354+
modelLibURLPrefix +
355+
modelVersion +
356+
"/Llama-3.2-1B-Instruct-q0f16-ctx4k_cs1k-webgpu.wasm",
357+
vram_required_MB: 2573.13,
358+
low_resource_required: true,
359+
overrides: {
360+
context_window_size: 4096,
361+
},
362+
},
363+
{
364+
model: "https://huggingface.co/mlc-ai/Llama-3.2-3B-Instruct-q4f32_1-MLC",
365+
model_id: "Llama-3.2-3B-Instruct-q4f32_1-MLC",
366+
model_lib:
367+
modelLibURLPrefix +
368+
modelVersion +
369+
"/Llama-3.2-3B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm",
370+
vram_required_MB: 2951.51,
371+
low_resource_required: true,
372+
overrides: {
373+
context_window_size: 4096,
374+
},
375+
},
376+
{
377+
model: "https://huggingface.co/mlc-ai/Llama-3.2-3B-Instruct-q4f16_1-MLC",
378+
model_id: "Llama-3.2-3B-Instruct-q4f16_1-MLC",
379+
model_lib:
380+
modelLibURLPrefix +
381+
modelVersion +
382+
"/Llama-3.2-3B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm",
383+
vram_required_MB: 2263.69,
384+
low_resource_required: true,
385+
overrides: {
386+
context_window_size: 4096,
387+
},
388+
},
389+
// Llama-3.1
312390
{
313391
model: "https://huggingface.co/mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC",
314392
model_id: "Llama-3.1-8B-Instruct-q4f32_1-MLC-1k",

0 commit comments

Comments
 (0)