Skip to content

Commit 8e28665

Browse files
authored
don't change output [no ci]
1 parent 0c2aad7 commit 8e28665

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

src/llama-context.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1276,7 +1276,7 @@ uint32_t llama_context::output_reserve(int32_t n_outputs) {
12761276

12771277
const auto n_batch = cparams.n_batch;
12781278
const auto n_vocab = vocab.n_tokens();
1279-
const auto n_embd = hparams.n_embd_full;
1279+
const auto n_embd = hparams.n_embd;
12801280

12811281
bool has_logits = true;
12821282
bool has_embd = cparams.embeddings;
@@ -1340,7 +1340,7 @@ uint32_t llama_context::output_reserve(int32_t n_outputs) {
13401340

13411341
void llama_context::output_reorder() {
13421342
const uint64_t n_vocab = model.vocab.n_tokens();
1343-
const uint64_t n_embd = model.hparams.n_embd_full;
1343+
const uint64_t n_embd = model.hparams.n_embd;
13441344

13451345
for (size_t s = 0; s < output_swaps.size(); ++s) {
13461346
const uint64_t i0 = output_swaps[s].i0;
@@ -1883,7 +1883,7 @@ size_t llama_context::state_write_data(llama_io_write_i & io) {
18831883
{
18841884
LLAMA_LOG_DEBUG("%s: - writing embeddings\n", __func__);
18851885

1886-
const uint64_t embd_size = std::min((uint64_t) this->embd_size, (uint64_t) n_outputs * model.hparams.n_embd_full);
1886+
const uint64_t embd_size = std::min((uint64_t) this->embd_size, (uint64_t) n_outputs * model.hparams.n_embd);
18871887

18881888
io.write(&embd_size, sizeof(embd_size));
18891889

0 commit comments

Comments
 (0)