File tree Expand file tree Collapse file tree 1 file changed +14
-14
lines changed Expand file tree Collapse file tree 1 file changed +14
-14
lines changed Original file line number Diff line number Diff line change @@ -3423,20 +3423,20 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
34233423 }
34243424 ).set_examples ({LLAMA_EXAMPLE_SERVER}));
34253425
3426- add_opt (common_arg (
3427- { " --fim-falcon-7b-spec " },
3428- string_format (" use quantized Falcon 7B model (note: can download weights from the internet)" ),
3429- [](common_params & params) {
3430- params.model .hf_repo = " maddes8cht/tiiuae-falcon-7b-gguf " ;
3431- params.model .hf_file = " tiiuae-falcon-7b-Q6_K .gguf" ;
3432- params.n_gpu_layers = 40 ;
3433- params.flash_attn = true ;
3434- params.n_ubatch = 1024 ;
3435- params.n_batch = 1024 ;
3436- params.n_ctx = 2048 ;
3437- params.n_cache_reuse = 256 ;
3438- }
3439- ).set_examples ({LLAMA_EXAMPLE_SERVER}));
3426+ add_opt (common_arg
3427+ ({ " --chat-smollm-3b-default " },
3428+ string_format (" use default SmolLM-v3 3B model (note: can download weights from the internet)" ),
3429+ [](common_params & params) {
3430+ params.model .hf_repo = " ggml-org/SmolLM3-3B-GGUF " ;
3431+ params.model .hf_file = " SmolLM3-Q8_0 .gguf" ;
3432+ params.n_gpu_layers = 40 ;
3433+ params.flash_attn = true ;
3434+ params.n_ubatch = 1024 ;
3435+ params.n_batch = 1024 ;
3436+ params.n_ctx = 65536 ;
3437+ params.n_cache_reuse = 256 ;
3438+ }
3439+ ).set_examples ({ LLAMA_EXAMPLE_SERVER }));
34403440
34413441
34423442 return ctx_arg;
You can’t perform that action at this time.
0 commit comments