Skip to content

Commit 2c2fbbd

Browse files
committed
cont : update comments [no ci]
1 parent d6d5e95 commit 2c2fbbd

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

src/llama-hparams.h

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ struct llama_hparams {
4141
uint32_t n_embd;
4242
uint32_t n_embd_features = 0;
4343
uint32_t n_layer;
44-
int32_t n_layer_kv_from_start = -1;
44+
int32_t n_layer_kv_from_start = -1; // if non-negative, the first n_layer_kv_from_start layers have KV cache
4545
uint32_t n_rot;
4646
uint32_t n_embd_head_k; // dimension of keys (d_k). d_q is assumed to be the same, but there are n_head q heads, and only n_head_kv k-v heads
4747
uint32_t n_embd_head_v; // dimension of values (d_v) aka n_embd_head

src/llama-memory.h

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -68,8 +68,8 @@ struct llama_memory_i {
6868
// this callback is used to filter out layers that should not be included in the cache
6969
using layer_filter_cb = std::function<bool(int32_t il)>;
7070

71-
// this callback is used to specify which layers should reuse KV cache from other layers
72-
// return negative value to indicate that the layer il should not reuse KV cache
71+
// this callback is used to specify which layers should reuse memory from other layers
72+
// return negative value to indicate that the layer il should not reuse memory
7373
using layer_reuse_cb = std::function<int32_t(int32_t il)>;
7474

7575
virtual ~llama_memory_i() = default;

0 commit comments

Comments
 (0)