@@ -73,17 +73,17 @@ static bool ggml_callback(struct ggml_tensor * t, bool ask, void * user_data) {
7373int main (int argc, char ** argv) {
7474 common_params params;
7575
76- params.model = " sesame-csm-backbone.gguf" ;
77- params.vocoder .model = " kyutai-mimi.gguf" ;
78- params.out_file = " output.wav" ;
79- params.prompt = " " ;
80- params.n_predict = 2048 ; // CSM's max trained seq length
81- params.sampling .top_k = 50 ; // default param from CSM python code
82- params.sampling .temp = 0.9 ; // default param from CSM python code
76+ params.model . path = " sesame-csm-backbone.gguf" ;
77+ params.vocoder .model . path = " kyutai-mimi.gguf" ;
78+ params.out_file = " output.wav" ;
79+ params.prompt = " " ;
80+ params.n_predict = 2048 ; // CSM's max trained seq length
81+ params.sampling .top_k = 50 ; // default param from CSM python code
82+ params.sampling .temp = 0.9 ; // default param from CSM python code
8383
8484 // HF model
85- params.model_url = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf" ;
86- params.vocoder .model_url = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf" ;
85+ params.model . url = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf" ;
86+ params.vocoder .model . url = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf" ;
8787
8888 if (!common_params_parse (argc, argv, params, LLAMA_EXAMPLE_TTS, print_usage)) {
8989 return 1 ;
@@ -104,10 +104,8 @@ int main(int argc, char ** argv) {
104104
105105 common_params params_decoder (params); // duplicate the params
106106 params_decoder.n_ctx = 64 ; // we never use more than this
107- string_replace_all (params_decoder.model , " -backbone" , " -decoder" );
108- if (!params_decoder.model_url .empty ()) {
109- string_replace_all (params_decoder.model_url , " -backbone" , " -decoder" );
110- }
107+ string_replace_all (params_decoder.model .path , " -backbone" , " -decoder" );
108+ string_replace_all (params_decoder.model .url , " -backbone" , " -decoder" );
111109
112110 common_init_result llama_backbone = common_init_from_params (params);
113111 llama_model * model_bb = llama_backbone.model .get ();
@@ -125,7 +123,7 @@ int main(int argc, char ** argv) {
125123 return ENOENT;
126124 }
127125
128- mimi_model mimi (params.vocoder .model .c_str (), true );
126+ mimi_model mimi (params.vocoder .model .path . c_str (), true );
129127
130128 // tokenize the prompt
131129 const llama_vocab * vocab = llama_model_get_vocab (model_bb);
0 commit comments