@@ -308,7 +308,85 @@ export const functionCallingModelIds = [
308
308
export const prebuiltAppConfig : AppConfig = {
309
309
useIndexedDBCache : false ,
310
310
model_list : [
311
- // Llama-3
311
+ {
312
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q4f32_1-MLC" ,
313
+ model_id : "Llama-3.2-1B-Instruct-q4f32_1-MLC" ,
314
+ model_lib :
315
+ modelLibURLPrefix +
316
+ modelVersion +
317
+ "/Llama-3.2-1B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm" ,
318
+ vram_required_MB : 1128.82 ,
319
+ low_resource_required : true ,
320
+ overrides : {
321
+ context_window_size : 4096 ,
322
+ } ,
323
+ } ,
324
+ {
325
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q4f16_1-MLC" ,
326
+ model_id : "Llama-3.2-1B-Instruct-q4f16_1-MLC" ,
327
+ model_lib :
328
+ modelLibURLPrefix +
329
+ modelVersion +
330
+ "/Llama-3.2-1B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm" ,
331
+ vram_required_MB : 879.04 ,
332
+ low_resource_required : true ,
333
+ overrides : {
334
+ context_window_size : 4096 ,
335
+ } ,
336
+ } ,
337
+ {
338
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q0f32-MLC" ,
339
+ model_id : "Llama-3.2-1B-Instruct-q0f32-MLC" ,
340
+ model_lib :
341
+ modelLibURLPrefix +
342
+ modelVersion +
343
+ "/Llama-3.2-1B-Instruct-q0f32-ctx4k_cs1k-webgpu.wasm" ,
344
+ vram_required_MB : 5106.26 ,
345
+ low_resource_required : true ,
346
+ overrides : {
347
+ context_window_size : 4096 ,
348
+ } ,
349
+ } ,
350
+ {
351
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-1B-Instruct-q0f16-MLC" ,
352
+ model_id : "Llama-3.2-1B-Instruct-q0f16-MLC" ,
353
+ model_lib :
354
+ modelLibURLPrefix +
355
+ modelVersion +
356
+ "/Llama-3.2-1B-Instruct-q0f16-ctx4k_cs1k-webgpu.wasm" ,
357
+ vram_required_MB : 2573.13 ,
358
+ low_resource_required : true ,
359
+ overrides : {
360
+ context_window_size : 4096 ,
361
+ } ,
362
+ } ,
363
+ {
364
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-3B-Instruct-q4f32_1-MLC" ,
365
+ model_id : "Llama-3.2-3B-Instruct-q4f32_1-MLC" ,
366
+ model_lib :
367
+ modelLibURLPrefix +
368
+ modelVersion +
369
+ "/Llama-3.2-3B-Instruct-q4f32_1-ctx4k_cs1k-webgpu.wasm" ,
370
+ vram_required_MB : 2951.51 ,
371
+ low_resource_required : true ,
372
+ overrides : {
373
+ context_window_size : 4096 ,
374
+ } ,
375
+ } ,
376
+ {
377
+ model : "https://huggingface.co/mlc-ai/Llama-3.2-3B-Instruct-q4f16_1-MLC" ,
378
+ model_id : "Llama-3.2-3B-Instruct-q4f16_1-MLC" ,
379
+ model_lib :
380
+ modelLibURLPrefix +
381
+ modelVersion +
382
+ "/Llama-3.2-3B-Instruct-q4f16_1-ctx4k_cs1k-webgpu.wasm" ,
383
+ vram_required_MB : 2263.69 ,
384
+ low_resource_required : true ,
385
+ overrides : {
386
+ context_window_size : 4096 ,
387
+ } ,
388
+ } ,
389
+ // Llama-3.1
312
390
{
313
391
model : "https://huggingface.co/mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC" ,
314
392
model_id : "Llama-3.1-8B-Instruct-q4f32_1-MLC-1k" ,
0 commit comments