@@ -1394,7 +1394,9 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
13941394 options.push_back ({ " *" , " --keep N" , " number of tokens to keep from the initial prompt (default: %d, -1 = all)" , params.n_keep });
13951395 options.push_back ({ " *" , " --chunks N" , " max number of chunks to process (default: %d, -1 = all)" , params.n_chunks });
13961396 options.push_back ({ " *" , " -fa, --flash-attn" , " enable Flash Attention (default: %s)" , params.flash_attn ? " enabled" : " disabled" });
1397- options.push_back ({ " *" , " -p, --prompt PROMPT" , " prompt to start generation with (default: '%s')" , params.prompt .c_str () });
1397+ options.push_back ({ " *" , " -p, --prompt PROMPT" , " prompt to start generation with\n "
1398+ " in conversation mode, this will be used as system prompt\n "
1399+ " (default: '%s')" , params.prompt .c_str () });
13981400 options.push_back ({ " *" , " -f, --file FNAME" , " a file containing the prompt (default: none)" });
13991401 options.push_back ({ " *" , " --in-file FNAME" , " an input file (repeat to specify multiple files)" });
14001402 options.push_back ({ " *" , " -bf, --binary-file FNAME" , " binary file containing the prompt (default: none)" });
@@ -1409,7 +1411,9 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
14091411 " halt generation at PROMPT, return control in interactive mode\n "
14101412 " can be specified more than once for multiple prompts" });
14111413 options.push_back ({ " main" , " -sp, --special" , " special tokens output enabled (default: %s)" , params.special ? " true" : " false" });
1412- options.push_back ({ " main" , " -cnv, --conversation" , " run in conversation mode (does not print special tokens and suffix/prefix, use default chat template) (default: %s)" , params.conversation ? " true" : " false" });
1414+ options.push_back ({ " main" , " -cnv, --conversation" , " run in conversation mode, does not print special tokens and suffix/prefix\n "
1415+ " if suffix/prefix are not specified, default chat template will be used\n "
1416+ " (default: %s)" , params.conversation ? " true" : " false" });
14131417 options.push_back ({ " main infill" , " -i, --interactive" , " run in interactive mode (default: %s)" , params.interactive ? " true" : " false" });
14141418 options.push_back ({ " main infill" , " -if, --interactive-first" , " run in interactive mode and wait for input right away (default: %s)" , params.interactive_first ? " true" : " false" });
14151419 options.push_back ({ " main infill" , " -mli, --multiline-input" , " allows you to write or paste multiple lines without ending each in '\\ '" });
@@ -1453,6 +1457,7 @@ void gpt_params_print_usage(int /*argc*/, char ** argv, const gpt_params & param
14531457 options.push_back ({ " main" , " --cfg-scale N" , " strength of guidance (default: %.1f, 1.0 = disable)" , (double )sparams.cfg_scale });
14541458 options.push_back ({ " main" , " --chat-template JINJA_TEMPLATE" ,
14551459 " set custom jinja chat template (default: template taken from model's metadata)\n "
1460+ " if suffix/prefix are specified, template will be disabled\n "
14561461 " only commonly used templates are accepted:\n "
14571462 " https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template" });
14581463 options.push_back ({ " grammar" });
0 commit comments