@@ -1467,6 +1467,14 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
14671467 params.swa_full = true ;
14681468 }
14691469 ).set_env (" LLAMA_ARG_SWA_FULL" ));
1470+ add_opt (common_arg (
1471+ {" --kv-unified" , " -kvu" },
1472+ string_format (" use single unified KV buffer for the KV cache of all sequences (default: %s)\n "
1473+ " [(more info)](https://github.com/ggml-org/llama.cpp/pull/14363)" , params.kv_unified ? " true" : " false" ),
1474+ [](common_params & params) {
1475+ params.kv_unified = true ;
1476+ }
1477+ ).set_env (" LLAMA_ARG_KV_SPLIT" ));
14701478 add_opt (common_arg (
14711479 {" --no-context-shift" },
14721480 string_format (" disables context shift on infinite text generation (default: %s)" , params.ctx_shift ? " disabled" : " enabled" ),
@@ -1607,7 +1615,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
16071615 [](common_params & params, const std::string & value) {
16081616 params.antiprompt .emplace_back (value);
16091617 }
1610- ).set_examples ({LLAMA_EXAMPLE_MAIN}));
1618+ ).set_examples ({LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_SERVER }));
16111619 add_opt (common_arg (
16121620 {" -sp" , " --special" },
16131621 string_format (" special tokens output enabled (default: %s)" , params.special ? " true" : " false" ),
@@ -2650,6 +2658,13 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
26502658 params.i_chunk = value;
26512659 }
26522660 ).set_examples ({LLAMA_EXAMPLE_IMATRIX}));
2661+ add_opt (common_arg (
2662+ {" --show-statistics" },
2663+ string_format (" show imatrix statistics and then exit (default: %s)" , params.show_statistics ? " true" : " false" ),
2664+ [](common_params & params) {
2665+ params.show_statistics = true ;
2666+ }
2667+ ).set_examples ({LLAMA_EXAMPLE_IMATRIX}));
26532668 add_opt (common_arg (
26542669 {" --parse-special" },
26552670 string_format (" prase special tokens (chat, tool, etc) (default: %s)" , params.parse_special ? " true" : " false" ),
@@ -3426,5 +3441,34 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
34263441 }
34273442 ).set_examples ({LLAMA_EXAMPLE_SERVER}));
34283443
3444+ // diffusion parameters
3445+ add_opt (common_arg (
3446+ { " --diffusion-steps" }, " N" ,
3447+ string_format (" number of diffusion steps (default: %d)" , params.diffusion .steps ),
3448+ [](common_params & params, int value) { params.diffusion .steps = value; }
3449+ ).set_examples ({ LLAMA_EXAMPLE_DIFFUSION }));
3450+ add_opt (common_arg (
3451+ { " --diffusion-eps" }, " F" ,
3452+ string_format (" epsilon for timesteps (default: %.6f)" , (double ) params.diffusion .eps ),
3453+ [](common_params & params, const std::string & value) { params.diffusion .eps = std::stof (value); }
3454+ ).set_examples ({ LLAMA_EXAMPLE_DIFFUSION }));
3455+ add_opt (common_arg (
3456+ { " --diffusion-algorithm" }, " N" ,
3457+ string_format (" diffusion algorithm: 0=ORIGIN, 1=MASKGIT_PLUS, 2=TOPK_MARGIN, 3=ENTROPY (default: %d)" ,
3458+ params.diffusion .algorithm ),
3459+ [](common_params & params, int value) { params.diffusion .algorithm = value; }
3460+ ).set_examples ({ LLAMA_EXAMPLE_DIFFUSION }));
3461+ add_opt (common_arg (
3462+ { " --diffusion-alg-temp" }, " F" ,
3463+ string_format (" algorithm temperature (default: %.3f)" , (double ) params.diffusion .alg_temp ),
3464+ [](common_params & params, const std::string & value) { params.diffusion .alg_temp = std::stof (value); }
3465+ ).set_examples ({ LLAMA_EXAMPLE_DIFFUSION }));
3466+ add_opt (common_arg (
3467+ { " --diffusion-visual" },
3468+ string_format (" enable visual diffusion mode (show progressive generation) (default: %s)" ,
3469+ params.diffusion .visual_mode ? " true" : " false" ),
3470+ [](common_params & params) { params.diffusion .visual_mode = true ; }
3471+ ).set_examples ({ LLAMA_EXAMPLE_DIFFUSION }));
3472+
34293473 return ctx_arg;
34303474}
0 commit comments