Skip to content

Commit 18842b6

Browse files
committed
add TODO
1 parent f706358 commit 18842b6

File tree

1 file changed

+5
-0
lines changed

1 file changed

+5
-0
lines changed

src/llama-kv-cache.cpp

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1579,6 +1579,9 @@ void llama_kv_cache::state_write_meta(llama_io_write_i & io, const cell_ranges_t
15791579
io.write(&pos, sizeof(pos));
15801580
io.write(&n_seq_id, sizeof(n_seq_id));
15811581

1582+
// TODO: we also need to save llama_kv_cell_ext when apply_ubatch() support loading it
1583+
// see: https://github.com/ggml-org/llama.cpp/pull/16825#issuecomment-3460868350
1584+
15821585
for (const auto & seq_id : seq_ids) {
15831586
io.write(&seq_id, sizeof(seq_id));
15841587
}
@@ -1724,6 +1727,8 @@ bool llama_kv_cache::state_read_meta(llama_io_read_i & io, uint32_t strm, uint32
17241727
return false;
17251728
}
17261729

1730+
// TODO: we cannot yet restore llama_kv_cell_ext as the apply_ubatch() does not support it yet
1731+
// see: https://github.com/ggml-org/llama.cpp/pull/16825#issuecomment-3460868350
17271732
apply_ubatch(sinfo, ubatch);
17281733

17291734
const auto head_cur = sinfo.head();

0 commit comments

Comments
 (0)