You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
LOG_INF("%s: Auto-determined training context size (n_ctx_train): %d\n", __func__, n_ctx_train);
134
-
if (n_ctx_train > llama_n_ctx(ctx)) {
135
-
LOG_DBG("%s: Auto-determined training context size (%d) is larger than model's context size (%d). Sequences will be truncated.\n", __func__, n_ctx_train, llama_n_ctx(ctx));
133
+
effective_n_ctx_train = max_seq_len_in_dataset;
134
+
LOG_INF("%s: Auto-determined training context size (n_ctx_train): %d\n", __func__, effective_n_ctx_train);
135
+
if (effective_n_ctx_train > llama_model_n_ctx_train(model)) {
136
+
LOG_DBG("%s: Auto-determined training context size (%d) is larger than model's native context size (%d). Sequences will be truncated by llama_opt_dataset_add_data.\n", __func__, effective_n_ctx_train, llama_model_n_ctx_train(model));
137
+
}
138
+
} else {
139
+
LOG_INF("%s: Using user-specified training context size (n_ctx_train): %d\n", __func__, effective_n_ctx_train);
140
+
if (effective_n_ctx_train > llama_model_n_ctx_train(model)) {
141
+
LOG_DBG("%s: User-specified training context size (%d) is larger than model's native context size (%d). Sequences will be truncated by llama_opt_dataset_add_data.\n", __func__, effective_n_ctx_train, llama_model_n_ctx_train(model));
0 commit comments