@@ -21,14 +21,12 @@ static std::string format_input_text(const std::string & prompt, bool use_chat_t
2121 auto chat_templates = common_chat_templates_init (model, " " );
2222
2323 common_chat_templates_inputs inputs;
24- // Create user message
2524 common_chat_msg user_msg;
2625 user_msg.role = " user" ;
2726 user_msg.content = prompt;
2827 inputs.add_generation_prompt = true ;
2928 inputs.messages .push_back (user_msg);
3029
31- // Apply chat template
3230 auto result = common_chat_templates_apply (chat_templates.get (), inputs);
3331
3432 return result.prompt ;
@@ -90,12 +88,10 @@ int main(int argc, char ** argv) {
9088
9189 common_params params;
9290
93- // Use the standard common_params_parse with LLAMA_EXAMPLE_DIFFUSION
9491 if (!common_params_parse (argc, argv, params, LLAMA_EXAMPLE_DIFFUSION)) {
9592 return 1 ;
9693 }
9794
98- // Print diffusion parameters
9995 const char * alg_names[] = { " ORIGIN" , " MASKGIT_PLUS" , " TOPK_MARGIN" , " ENTROPY" };
10096 const char * alg_name = (params.diffusion .algorithm >= 0 && params.diffusion .algorithm <= 3 ) ?
10197 alg_names[params.diffusion .algorithm ] :
@@ -117,7 +113,6 @@ int main(int argc, char ** argv) {
117113 return 1 ;
118114 }
119115
120- // Create context
121116 llama_context_params ctx_params = llama_context_default_params ();
122117 ctx_params.n_ctx = params.n_ctx ;
123118 ctx_params.n_batch = params.n_batch ;
0 commit comments