Skip to content

Commit 29fd7b5

Browse files
committed
llama : chat
ggml-ci
1 parent c8669a0 commit 29fd7b5

File tree

6 files changed

+576
-583
lines changed

6 files changed

+576
-583
lines changed

src/CMakeLists.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -11,6 +11,7 @@ add_library(llama
1111
llama.cpp
1212
llama-arch.cpp
1313
llama-batch.cpp
14+
llama-chat.cpp
1415
llama-context.cpp
1516
llama-adapter.cpp
1617
llama-grammar.cpp

src/llama-arch.cpp

Lines changed: 0 additions & 47 deletions
Original file line numberDiff line numberDiff line change
@@ -1385,50 +1385,3 @@ llm_arch llm_arch_from_string(const std::string & name) {
13851385
const llm_tensor_info & llm_tensor_info_for(llm_tensor tensor) {
13861386
return LLM_TENSOR_INFOS.at(tensor);
13871387
}
1388-
1389-
// chat templates
1390-
1391-
static const std::map<std::string, llm_chat_template> LLM_CHAT_TEMPLATES = {
1392-
{ "chatml", LLM_CHAT_TEMPLATE_CHATML },
1393-
{ "llama2", LLM_CHAT_TEMPLATE_LLAMA_2 },
1394-
{ "llama2-sys", LLM_CHAT_TEMPLATE_LLAMA_2_SYS },
1395-
{ "llama2-sys-bos", LLM_CHAT_TEMPLATE_LLAMA_2_SYS_BOS },
1396-
{ "llama2-sys-strip", LLM_CHAT_TEMPLATE_LLAMA_2_SYS_STRIP },
1397-
{ "mistral-v1", LLM_CHAT_TEMPLATE_MISTRAL_V1 },
1398-
{ "mistral-v3", LLM_CHAT_TEMPLATE_MISTRAL_V3 },
1399-
{ "mistral-v3-tekken", LLM_CHAT_TEMPLATE_MISTRAL_V3_TEKKEN },
1400-
{ "mistral-v7", LLM_CHAT_TEMPLATE_MISTRAL_V7 },
1401-
{ "phi3", LLM_CHAT_TEMPLATE_PHI_3 },
1402-
{ "zephyr", LLM_CHAT_TEMPLATE_ZEPHYR },
1403-
{ "monarch", LLM_CHAT_TEMPLATE_MONARCH },
1404-
{ "gemma", LLM_CHAT_TEMPLATE_GEMMA },
1405-
{ "orion", LLM_CHAT_TEMPLATE_ORION },
1406-
{ "openchat", LLM_CHAT_TEMPLATE_OPENCHAT },
1407-
{ "vicuna", LLM_CHAT_TEMPLATE_VICUNA },
1408-
{ "vicuna-orca", LLM_CHAT_TEMPLATE_VICUNA_ORCA },
1409-
{ "deepseek", LLM_CHAT_TEMPLATE_DEEPSEEK },
1410-
{ "deepseek2", LLM_CHAT_TEMPLATE_DEEPSEEK_2 },
1411-
{ "command-r", LLM_CHAT_TEMPLATE_COMMAND_R },
1412-
{ "llama3", LLM_CHAT_TEMPLATE_LLAMA_3 },
1413-
{ "chatglm3", LLM_CHAT_TEMPLATE_CHATGML_3 },
1414-
{ "chatglm4", LLM_CHAT_TEMPLATE_CHATGML_4 },
1415-
{ "minicpm", LLM_CHAT_TEMPLATE_MINICPM },
1416-
{ "exaone3", LLM_CHAT_TEMPLATE_EXAONE_3 },
1417-
{ "rwkv-world", LLM_CHAT_TEMPLATE_RWKV_WORLD },
1418-
{ "granite", LLM_CHAT_TEMPLATE_GRANITE },
1419-
{ "gigachat", LLM_CHAT_TEMPLATE_GIGACHAT },
1420-
};
1421-
1422-
llm_chat_template llm_chat_template_from_str(const std::string & name) {
1423-
return LLM_CHAT_TEMPLATES.at(name);
1424-
}
1425-
1426-
int32_t llama_chat_builtin_templates(const char ** output, size_t len) {
1427-
auto it = LLM_CHAT_TEMPLATES.begin();
1428-
for (size_t i = 0; i < std::min(len, LLM_CHAT_TEMPLATES.size()); i++) {
1429-
output[i] = it->first.c_str();
1430-
std::advance(it, 1);
1431-
}
1432-
return (int32_t) LLM_CHAT_TEMPLATES.size();
1433-
}
1434-

src/llama-arch.h

Lines changed: 0 additions & 39 deletions
Original file line numberDiff line numberDiff line change
@@ -388,42 +388,3 @@ const char * llm_arch_name(llm_arch arch);
388388
llm_arch llm_arch_from_string(const std::string & name);
389389

390390
const llm_tensor_info & llm_tensor_info_for(llm_tensor tensor);
391-
392-
//
393-
// Chat templates
394-
// TODO: maybe move these to a separate module
395-
//
396-
397-
enum llm_chat_template {
398-
LLM_CHAT_TEMPLATE_CHATML,
399-
LLM_CHAT_TEMPLATE_LLAMA_2,
400-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS,
401-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_BOS,
402-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_STRIP,
403-
LLM_CHAT_TEMPLATE_MISTRAL_V1,
404-
LLM_CHAT_TEMPLATE_MISTRAL_V3,
405-
LLM_CHAT_TEMPLATE_MISTRAL_V3_TEKKEN,
406-
LLM_CHAT_TEMPLATE_MISTRAL_V7,
407-
LLM_CHAT_TEMPLATE_PHI_3,
408-
LLM_CHAT_TEMPLATE_ZEPHYR,
409-
LLM_CHAT_TEMPLATE_MONARCH,
410-
LLM_CHAT_TEMPLATE_GEMMA,
411-
LLM_CHAT_TEMPLATE_ORION,
412-
LLM_CHAT_TEMPLATE_OPENCHAT,
413-
LLM_CHAT_TEMPLATE_VICUNA,
414-
LLM_CHAT_TEMPLATE_VICUNA_ORCA,
415-
LLM_CHAT_TEMPLATE_DEEPSEEK,
416-
LLM_CHAT_TEMPLATE_DEEPSEEK_2,
417-
LLM_CHAT_TEMPLATE_COMMAND_R,
418-
LLM_CHAT_TEMPLATE_LLAMA_3,
419-
LLM_CHAT_TEMPLATE_CHATGML_3,
420-
LLM_CHAT_TEMPLATE_CHATGML_4,
421-
LLM_CHAT_TEMPLATE_MINICPM,
422-
LLM_CHAT_TEMPLATE_EXAONE_3,
423-
LLM_CHAT_TEMPLATE_RWKV_WORLD,
424-
LLM_CHAT_TEMPLATE_GRANITE,
425-
LLM_CHAT_TEMPLATE_GIGACHAT,
426-
LLM_CHAT_TEMPLATE_UNKNOWN,
427-
};
428-
429-
llm_chat_template llm_chat_template_from_str(const std::string & name);

0 commit comments

Comments
 (0)