@@ -1760,7 +1760,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
17601760 ).set_examples ({LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_SPECULATIVE, LLAMA_EXAMPLE_LOOKUP}));
17611761 add_opt (common_arg (
17621762 {" -t" , " --threads" }, " N" ,
1763- string_format (" number of threads to use during generation (default: %d)" , params.cpuparams .n_threads ),
1763+ string_format (" number of CPU threads to use during generation (default: %d)" , params.cpuparams .n_threads ),
17641764 [](common_params & params, int value) {
17651765 params.cpuparams .n_threads = value;
17661766 if (params.cpuparams .n_threads <= 0 ) {
@@ -2030,7 +2030,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
20302030 params.system_prompt .pop_back ();
20312031 }
20322032 }
2033- ).set_examples ({LLAMA_EXAMPLE_MAIN}));
2033+ ).set_examples ({LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_DIFFUSION }));
20342034 add_opt (common_arg (
20352035 {" --in-file" }, " FNAME" ,
20362036 " an input file (repeat to specify multiple files)" ,
@@ -2768,6 +2768,20 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
27682768 params.image .emplace_back (value);
27692769 }
27702770 ).set_examples ({LLAMA_EXAMPLE_MTMD}));
2771+ add_opt (common_arg (
2772+ {" --image-min-tokens" }, " N" ,
2773+ " minimum number of tokens each image can take, only used by vision models with dynamic resolution (default: read from model)" ,
2774+ [](common_params & params, int value) {
2775+ params.image_min_tokens = value;
2776+ }
2777+ ).set_examples (mmproj_examples).set_env (" LLAMA_ARG_IMAGE_MIN_TOKENS" ));
2778+ add_opt (common_arg (
2779+ {" --image-max-tokens" }, " N" ,
2780+ " maximum number of tokens each image can take, only used by vision models with dynamic resolution (default: read from model)" ,
2781+ [](common_params & params, int value) {
2782+ params.image_max_tokens = value;
2783+ }
2784+ ).set_examples (mmproj_examples).set_env (" LLAMA_ARG_IMAGE_MAX_TOKENS" ));
27712785 if (llama_supports_rpc ()) {
27722786 add_opt (common_arg (
27732787 {" --rpc" }, " SERVERS" ,
@@ -3203,7 +3217,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
32033217 ).set_examples ({LLAMA_EXAMPLE_IMATRIX}));
32043218 add_opt (common_arg (
32053219 {" --parse-special" },
3206- string_format (" prase special tokens (chat, tool, etc) (default: %s)" , params.parse_special ? " true" : " false" ),
3220+ string_format (" parse special tokens (chat, tool, etc) (default: %s)" , params.parse_special ? " true" : " false" ),
32073221 [](common_params & params) {
32083222 params.parse_special = true ;
32093223 }
@@ -3248,7 +3262,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
32483262 ).set_examples ({LLAMA_EXAMPLE_EMBEDDING}));
32493263 add_opt (common_arg (
32503264 {" --embd-output-format" }, " FORMAT" ,
3251- " empty = default, \" array\" = [[],[]...], \" json\" = openai style, \" json+\" = same \" json\" + cosine similarity matrix" ,
3265+ " empty = default, \" array\" = [[],[]...], \" json\" = openai style, \" json+\" = same \" json\" + cosine similarity matrix, \" raw \" = plain whitespace-delimited output (one embedding per line) " ,
32523266 [](common_params & params, const std::string & value) {
32533267 params.embd_out = value;
32543268 }
@@ -3435,7 +3449,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
34353449 [](common_params & params) {
34363450 params.use_jinja = true ;
34373451 }
3438- ).set_examples ({LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_MAIN}).set_env (" LLAMA_ARG_JINJA" ));
3452+ ).set_examples ({LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_MTMD }).set_env (" LLAMA_ARG_JINJA" ));
34393453 add_opt (common_arg (
34403454 {" --reasoning-format" }, " FORMAT" ,
34413455 " controls whether thought tags are allowed and/or extracted from the response, and in which format they're returned; one of:\n "
0 commit comments