Skip to content

Commit ca5074d

Browse files
committed
Merge branch 'eso_b6267' into crokeso
2 parents 31cf666 + 3014fa1 commit ca5074d

File tree

115 files changed

+3176
-31641
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

115 files changed

+3176
-31641
lines changed

.github/workflows/build-riscv-native.yml

Lines changed: 0 additions & 43 deletions
This file was deleted.

Makefile

Lines changed: 49 additions & 47 deletions
Large diffs are not rendered by default.

colab.ipynb

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@
7070
"WCommand = \"\"\n",
7171
"#@markdown <hr>\n",
7272
"LoadTTSModel = False #@param {type:\"boolean\"}\n",
73-
"TTSModel = \"https://huggingface.co/koboldcpp/tts/resolve/main/OuteTTS-0.2-500M-Q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/tts/resolve/main/OuteTTS-0.2-500M-Q4_0.gguf\"]{allow-input: true}\n",
73+
"TTSModel = \"https://huggingface.co/koboldcpp/tts/resolve/main/OuteTTS-0.2-500M-Q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/tts/resolve/main/OuteTTS-0.2-500M-Q4_0.gguf\",\"https://huggingface.co/koboldcpp/tts/resolve/main/Kokoro_no_espeak_Q4.gguf\"]{allow-input: true}\n",
7474
"WavTokModel = \"https://huggingface.co/koboldcpp/tts/resolve/main/WavTokenizer-Large-75-Q4_0.gguf\" #@param [\"https://huggingface.co/koboldcpp/tts/resolve/main/WavTokenizer-Large-75-Q4_0.gguf\"]{allow-input: true}\n",
7575
"TTSCommand = \"\"\n",
7676
"#@markdown <hr>\n",
@@ -127,6 +127,8 @@
127127
" WCommand = \"\"\n",
128128
"if TTSModel and WavTokModel and LoadTTSModel:\n",
129129
" TTSCommand = f\"--ttsmodel {TTSModel} --ttswavtokenizer {WavTokModel} --ttsgpu\"\n",
130+
"elif TTSModel and LoadTTSModel:\n",
131+
" TTSCommand = f\"--ttsmodel {TTSModel} --ttsgpu\"\n",
130132
"else:\n",
131133
" TTSCommand = \"\"\n",
132134
"if EmbeddingsModel and LoadEmbeddingsModel:\n",

common/arg.cpp

Lines changed: 8 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1533,6 +1533,13 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
15331533
params.ctx_shift = false;
15341534
}
15351535
).set_examples({LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_IMATRIX, LLAMA_EXAMPLE_PERPLEXITY}).set_env("LLAMA_ARG_NO_CONTEXT_SHIFT"));
1536+
add_opt(common_arg(
1537+
{"--context-shift"},
1538+
string_format("enables context shift on infinite text generation (default: %s)", params.ctx_shift ? "enabled" : "disabled"),
1539+
[](common_params & params) {
1540+
params.ctx_shift = true;
1541+
}
1542+
).set_examples({LLAMA_EXAMPLE_MAIN, LLAMA_EXAMPLE_SERVER, LLAMA_EXAMPLE_IMATRIX, LLAMA_EXAMPLE_PERPLEXITY}).set_env("LLAMA_ARG_CONTEXT_SHIFT"));
15361543
add_opt(common_arg(
15371544
{"--chunks"}, "N",
15381545
string_format("max number of chunks to process (default: %d, -1 = all)", params.n_chunks),
@@ -1826,7 +1833,7 @@ common_params_context common_params_parser_init(common_params & params, llama_ex
18261833
[](common_params & params, const std::string & value) {
18271834
params.sampling.top_n_sigma = std::stof(value);
18281835
}
1829-
).set_examples({LLAMA_EXAMPLE_MAIN}).set_sparam());
1836+
).set_sparam());
18301837
add_opt(common_arg(
18311838
{"--xtc-probability"}, "N",
18321839
string_format("xtc probability (default: %.1f, 0.0 = disabled)", (double)params.sampling.xtc_probability),

common/common.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -366,7 +366,7 @@ struct common_params {
366366
bool cont_batching = true; // insert new sequences for decoding on-the-fly
367367
bool flash_attn = false; // flash attention
368368
bool no_perf = false; // disable performance metrics
369-
bool ctx_shift = true; // context shift on inifinite text generation
369+
bool ctx_shift = false; // context shift on infinite text generation
370370
bool swa_full = false; // use full-size SWA cache (https://github.com/ggml-org/llama.cpp/pull/13194#issuecomment-2868343055)
371371
bool kv_unified = false; // enable unified KV cache
372372

0 commit comments

Comments
 (0)