@@ -73,17 +73,17 @@ static bool ggml_callback(struct ggml_tensor * t, bool ask, void * user_data) {
7373int  main (int  argc, char  ** argv) {
7474    common_params params;
7575
76-     params.model            = " sesame-csm-backbone.gguf"  ;
77-     params.vocoder .model    = " kyutai-mimi.gguf"  ;
78-     params.out_file        = " output.wav"  ;
79-     params.prompt          = " "  ;
80-     params.n_predict       = 2048 ; //  CSM's max trained seq length
81-     params.sampling .top_k  = 50 ;   //  default param from CSM python code
82-     params.sampling .temp   = 0.9 ;  //  default param from CSM python code
76+     params.model . path          = " sesame-csm-backbone.gguf"  ;
77+     params.vocoder .model . path  = " kyutai-mimi.gguf"  ;
78+     params.out_file             = " output.wav"  ;
79+     params.prompt               = " "  ;
80+     params.n_predict            = 2048 ; //  CSM's max trained seq length
81+     params.sampling .top_k       = 50 ;   //  default param from CSM python code
82+     params.sampling .temp        = 0.9 ;  //  default param from CSM python code
8383
8484    //  HF model
85-     params.model_url          = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf"  ;
86-     params.vocoder .model_url  = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf"  ;
85+     params.model . url          = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/sesame-csm-backbone.gguf"  ;
86+     params.vocoder .model . url  = " https://huggingface.co/ggml-org/sesame-csm-1b-GGUF/resolve/main/kyutai-mimi.gguf"  ;
8787
8888    if  (!common_params_parse (argc, argv, params, LLAMA_EXAMPLE_TTS, print_usage)) {
8989        return  1 ;
@@ -104,10 +104,8 @@ int main(int argc, char ** argv) {
104104
105105    common_params params_decoder (params); //  duplicate the params
106106    params_decoder.n_ctx  = 64 ; //  we never use more than this
107-     string_replace_all (params_decoder.model , " -backbone"  , " -decoder"  );
108-     if  (!params_decoder.model_url .empty ()) {
109-         string_replace_all (params_decoder.model_url , " -backbone"  , " -decoder"  );
110-     }
107+     string_replace_all (params_decoder.model .path , " -backbone"  , " -decoder"  );
108+     string_replace_all (params_decoder.model .url ,  " -backbone"  , " -decoder"  );
111109
112110    common_init_result llama_backbone = common_init_from_params (params);
113111    llama_model   * model_bb = llama_backbone.model .get ();
@@ -125,7 +123,7 @@ int main(int argc, char ** argv) {
125123        return  ENOENT;
126124    }
127125
128-     mimi_model mimi (params.vocoder .model .c_str (), true );
126+     mimi_model mimi (params.vocoder .model .path . c_str (), true );
129127
130128    //  tokenize the prompt
131129    const  llama_vocab * vocab = llama_model_get_vocab (model_bb);
0 commit comments