@@ -1385,50 +1385,3 @@ llm_arch llm_arch_from_string(const std::string & name) {
13851385const llm_tensor_info & llm_tensor_info_for (llm_tensor tensor) {
13861386 return LLM_TENSOR_INFOS.at (tensor);
13871387}
1388-
1389- // chat templates
1390-
1391- static const std::map<std::string, llm_chat_template> LLM_CHAT_TEMPLATES = {
1392- { " chatml" , LLM_CHAT_TEMPLATE_CHATML },
1393- { " llama2" , LLM_CHAT_TEMPLATE_LLAMA_2 },
1394- { " llama2-sys" , LLM_CHAT_TEMPLATE_LLAMA_2_SYS },
1395- { " llama2-sys-bos" , LLM_CHAT_TEMPLATE_LLAMA_2_SYS_BOS },
1396- { " llama2-sys-strip" , LLM_CHAT_TEMPLATE_LLAMA_2_SYS_STRIP },
1397- { " mistral-v1" , LLM_CHAT_TEMPLATE_MISTRAL_V1 },
1398- { " mistral-v3" , LLM_CHAT_TEMPLATE_MISTRAL_V3 },
1399- { " mistral-v3-tekken" , LLM_CHAT_TEMPLATE_MISTRAL_V3_TEKKEN },
1400- { " mistral-v7" , LLM_CHAT_TEMPLATE_MISTRAL_V7 },
1401- { " phi3" , LLM_CHAT_TEMPLATE_PHI_3 },
1402- { " zephyr" , LLM_CHAT_TEMPLATE_ZEPHYR },
1403- { " monarch" , LLM_CHAT_TEMPLATE_MONARCH },
1404- { " gemma" , LLM_CHAT_TEMPLATE_GEMMA },
1405- { " orion" , LLM_CHAT_TEMPLATE_ORION },
1406- { " openchat" , LLM_CHAT_TEMPLATE_OPENCHAT },
1407- { " vicuna" , LLM_CHAT_TEMPLATE_VICUNA },
1408- { " vicuna-orca" , LLM_CHAT_TEMPLATE_VICUNA_ORCA },
1409- { " deepseek" , LLM_CHAT_TEMPLATE_DEEPSEEK },
1410- { " deepseek2" , LLM_CHAT_TEMPLATE_DEEPSEEK_2 },
1411- { " command-r" , LLM_CHAT_TEMPLATE_COMMAND_R },
1412- { " llama3" , LLM_CHAT_TEMPLATE_LLAMA_3 },
1413- { " chatglm3" , LLM_CHAT_TEMPLATE_CHATGML_3 },
1414- { " chatglm4" , LLM_CHAT_TEMPLATE_CHATGML_4 },
1415- { " minicpm" , LLM_CHAT_TEMPLATE_MINICPM },
1416- { " exaone3" , LLM_CHAT_TEMPLATE_EXAONE_3 },
1417- { " rwkv-world" , LLM_CHAT_TEMPLATE_RWKV_WORLD },
1418- { " granite" , LLM_CHAT_TEMPLATE_GRANITE },
1419- { " gigachat" , LLM_CHAT_TEMPLATE_GIGACHAT },
1420- };
1421-
1422- llm_chat_template llm_chat_template_from_str (const std::string & name) {
1423- return LLM_CHAT_TEMPLATES.at (name);
1424- }
1425-
1426- int32_t llama_chat_builtin_templates (const char ** output, size_t len) {
1427- auto it = LLM_CHAT_TEMPLATES.begin ();
1428- for (size_t i = 0 ; i < std::min (len, LLM_CHAT_TEMPLATES.size ()); i++) {
1429- output[i] = it->first .c_str ();
1430- std::advance (it, 1 );
1431- }
1432- return (int32_t ) LLM_CHAT_TEMPLATES.size ();
1433- }
1434-
0 commit comments