File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -1276,7 +1276,7 @@ uint32_t llama_context::output_reserve(int32_t n_outputs) {
12761276
12771277 const auto n_batch = cparams.n_batch ;
12781278 const auto n_vocab = vocab.n_tokens ();
1279- const auto n_embd = hparams.n_embd_full ;
1279+ const auto n_embd = hparams.n_embd ;
12801280
12811281 bool has_logits = true ;
12821282 bool has_embd = cparams.embeddings ;
@@ -1340,7 +1340,7 @@ uint32_t llama_context::output_reserve(int32_t n_outputs) {
13401340
13411341void llama_context::output_reorder () {
13421342 const uint64_t n_vocab = model.vocab .n_tokens ();
1343- const uint64_t n_embd = model.hparams .n_embd_full ;
1343+ const uint64_t n_embd = model.hparams .n_embd ;
13441344
13451345 for (size_t s = 0 ; s < output_swaps.size (); ++s) {
13461346 const uint64_t i0 = output_swaps[s].i0 ;
@@ -1883,7 +1883,7 @@ size_t llama_context::state_write_data(llama_io_write_i & io) {
18831883 {
18841884 LLAMA_LOG_DEBUG (" %s: - writing embeddings\n " , __func__);
18851885
1886- const uint64_t embd_size = std::min ((uint64_t ) this ->embd_size , (uint64_t ) n_outputs * model.hparams .n_embd_full );
1886+ const uint64_t embd_size = std::min ((uint64_t ) this ->embd_size , (uint64_t ) n_outputs * model.hparams .n_embd );
18871887
18881888 io.write (&embd_size, sizeof (embd_size));
18891889
You can’t perform that action at this time.
0 commit comments