@@ -1745,15 +1745,15 @@ struct server_context {
17451745 llama_chat_message chat[] = {{" user" , " test" }};
17461746
17471747 if (use_jinja) {
1748- auto templates = llama_chat_templates_from_model (model, " " );
1749- GGML_ASSERT (templates.default_template );
1748+ auto templates = common_chat_templates_from_model (model, " " );
1749+ GGML_ASSERT (templates.template_default );
17501750 try {
1751- templates.default_template ->apply ({{
1751+ templates.template_default ->apply ({{
17521752 {" role" , " user" },
17531753 {" content" , " test" },
17541754 }}, json (), true );
1755- if (templates.tool_use_template ) {
1756- templates.tool_use_template ->apply ({{
1755+ if (templates.template_tool_use ) {
1756+ templates.template_tool_use ->apply ({{
17571757 {" role" , " user" },
17581758 {" content" , " test" },
17591759 }}, json (), true );
@@ -3631,8 +3631,8 @@ int main(int argc, char ** argv) {
36313631 auto get_chat_templates = [&ctx_server, &chat_templates_mutex, &chat_templates]() -> const llama_chat_templates & {
36323632 std::lock_guard<std::mutex> lock (chat_templates_mutex);
36333633 if (!chat_templates) {
3634- chat_templates = llama_chat_templates_from_model (ctx_server.model , ctx_server.params_base .chat_template );
3635- GGML_ASSERT (chat_templates->default_template );
3634+ chat_templates = common_chat_templates_from_model (ctx_server.model , ctx_server.params_base .chat_template );
3635+ GGML_ASSERT (chat_templates->template_default );
36363636 }
36373637 return *chat_templates;
36383638 };
@@ -3644,11 +3644,11 @@ int main(int argc, char ** argv) {
36443644 { " default_generation_settings" , ctx_server.default_generation_settings_for_props },
36453645 { " total_slots" , ctx_server.params_base .n_parallel },
36463646 { " model_path" , ctx_server.params_base .model },
3647- { " chat_template" , templates.default_template ->source () },
3647+ { " chat_template" , templates.template_default ->source () },
36483648 { " build_info" , build_info },
36493649 };
3650- if (ctx_server.params_base .use_jinja && templates.tool_use_template ) {
3651- data[" chat_template_tool_use" ] = templates.tool_use_template ->source ();
3650+ if (ctx_server.params_base .use_jinja && templates.template_tool_use ) {
3651+ data[" chat_template_tool_use" ] = templates.template_tool_use ->source ();
36523652 }
36533653
36543654 res_ok (res, data);
@@ -3871,7 +3871,7 @@ int main(int argc, char ** argv) {
38713871
38723872 auto body = json::parse (req.body );
38733873 const auto & templates = get_chat_templates ();
3874- const auto & chat_template = body.contains (" tools" ) && templates.tool_use_template ? *templates.tool_use_template : *templates.default_template ;
3874+ const auto & chat_template = body.contains (" tools" ) && templates.template_tool_use ? *templates.template_tool_use : *templates.template_default ;
38753875 json data = oaicompat_completion_params_parse (body, chat_template, params.use_jinja );
38763876
38773877 return handle_completions_impl (
@@ -4290,8 +4290,8 @@ int main(int argc, char ** argv) {
42904290
42914291 // print sample chat example to make it clear which template is used
42924292 LOG_INF (" %s: chat template, chat_template: %s, example_format: '%s'\n " , __func__,
4293- get_chat_templates ().default_template ->source ().c_str (),
4294- common_chat_format_example (*get_chat_templates ().default_template , ctx_server.params_base .use_jinja ).c_str ());
4293+ get_chat_templates ().template_default ->source ().c_str (),
4294+ common_chat_format_example (*get_chat_templates ().template_default , ctx_server.params_base .use_jinja ).c_str ());
42954295
42964296 ctx_server.queue_tasks .on_new_task (std::bind (
42974297 &server_context::process_single_task, &ctx_server, std::placeholders::_1));
0 commit comments