@@ -899,15 +899,7 @@ std::string fs_get_cache_file(const std::string & filename) {
899
899
// Model utils
900
900
//
901
901
902
- struct common_init_result common_init_from_params (common_params & params) {
903
- common_init_result iparams;
904
- auto mparams = common_model_params_to_llama (params);
905
-
906
- llama_model * model = llama_model_load_from_file (params.model .path .c_str (), mparams);
907
- if (model == NULL ) {
908
- LOG_ERR (" %s: failed to load model '%s'\n " , __func__, params.model .path .c_str ());
909
- return iparams;
910
- }
902
+ struct common_init_result common_init_from_model_and_params (llama_model* model, common_init_result iparams, common_params & params) {
911
903
912
904
const llama_vocab * vocab = llama_model_get_vocab (model);
913
905
@@ -1068,6 +1060,19 @@ struct common_init_result common_init_from_params(common_params & params) {
1068
1060
return iparams;
1069
1061
}
1070
1062
1063
+ struct common_init_result common_init_from_params (common_params & params) {
1064
+ common_init_result iparams;
1065
+ auto mparams = common_model_params_to_llama (params);
1066
+
1067
+ llama_model * model = llama_model_load_from_file (params.model .path .c_str (), mparams);
1068
+ if (model == NULL ) {
1069
+ LOG_ERR (" %s: failed to load model '%s'\n " , __func__, params.model .path .c_str ());
1070
+ return iparams;
1071
+ }
1072
+
1073
+ return common_init_from_model_and_params (model, std::move (iparams), params);
1074
+ }
1075
+
1071
1076
std::string get_model_endpoint () {
1072
1077
const char * model_endpoint_env = getenv (" MODEL_ENDPOINT" );
1073
1078
// We still respect the use of environment-variable "HF_ENDPOINT" for backward-compatibility.
0 commit comments