Skip to content

Commit e9dc476

Browse files
committed
fix tts-csm
1 parent 156b528 commit e9dc476

File tree

1 file changed

+12
-14
lines changed

1 file changed

+12
-14
lines changed

examples/tts/tts-csm.cpp

Lines changed: 12 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -73,17 +73,17 @@ static bool ggml_callback(struct ggml_tensor * t, bool ask, void * user_data) {
7373
int main(int argc, char ** argv) {
7474
common_params params;
7575

76-
params.model = "sesame-csm-backbone.gguf";
77-
params.vocoder.model = "kyutai-mimi.gguf";
78-
params.out_file = "output.wav";
79-
params.prompt = "";
80-
params.n_predict = 2048; // CSM's max trained seq length
81-
params.sampling.top_k = 50; // default param from CSM python code
82-
params.sampling.temp = 0.9; // default param from CSM python code
76+
params.model.path = "sesame-csm-backbone.gguf";
77+
params.vocoder.model.path = "kyutai-mimi.gguf";
78+
params.out_file = "output.wav";
79+
params.prompt = "";
80+
params.n_predict = 2048; // CSM's max trained seq length
81+
params.sampling.top_k = 50; // default param from CSM python code
82+
params.sampling.temp = 0.9; // default param from CSM python code
8383

8484
// HF model
85-
params.model_url = "https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf";
86-
params.vocoder.model_url = "https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf";
85+
params.model.url = "https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf";
86+
params.vocoder.model.url = "https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf";
8787

8888
if (!common_params_parse(argc, argv, params, LLAMA_EXAMPLE_TTS, print_usage)) {
8989
return 1;
@@ -104,10 +104,8 @@ int main(int argc, char ** argv) {
104104

105105
common_params params_decoder(params); // duplicate the params
106106
params_decoder.n_ctx = 64; // we never use more than this
107-
string_replace_all(params_decoder.model, "-backbone", "-decoder");
108-
if (!params_decoder.model_url.empty()) {
109-
string_replace_all(params_decoder.model_url, "-backbone", "-decoder");
110-
}
107+
string_replace_all(params_decoder.model.path, "-backbone", "-decoder");
108+
string_replace_all(params_decoder.model.url, "-backbone", "-decoder");
111109

112110
common_init_result llama_backbone = common_init_from_params(params);
113111
llama_model * model_bb = llama_backbone.model.get();
@@ -125,7 +123,7 @@ int main(int argc, char ** argv) {
125123
return ENOENT;
126124
}
127125

128-
mimi_model mimi(params.vocoder.model.c_str(), true);
126+
mimi_model mimi(params.vocoder.model.path.c_str(), true);
129127

130128
// tokenize the prompt
131129
const llama_vocab * vocab = llama_model_get_vocab(model_bb);

0 commit comments

Comments
 (0)