Skip to content

Commit 88536f7

Browse files
committed
debug log
1 parent 6277a0e commit 88536f7

File tree

1 file changed

+7
-6
lines changed

1 file changed

+7
-6
lines changed

src/llama-model.cpp

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -2946,29 +2946,30 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29462946
buft = ggml_backend_dev_buffer_type(cpu_dev);
29472947
}
29482948

2949+
LLAMA_LOG_INFO("n_head_kv: %d, kv_lora_rank: %d, n_embd_head_qk_nope: %d\n", n_head_kv, kv_lora_rank, n_embd_head_qk_nope);
29492950
ggml_context * ctx = ctx_for_buft(buft);
29502951
layer.wk_b = ggml_new_tensor_2d(ctx,
29512952
wkv_b->type,
29522953
n_head_kv * kv_lora_rank,
29532954
n_embd_head_qk_nope
29542955
);
2955-
LLAMA_LOG_DEBUG("111\n", 0);
2956+
LLAMA_LOG_INFO("wk_b shape: [%d, %d]\n", layer.wk_b->ne[0], layer.wk_b->ne[1]);
29562957
{
29572958
float *src = (float *)wkv_b->data;
29582959
float *dst = (float *)layer.wk_b->data;
29592960
int src_stride = wkv_b->ne[0]; // 原始张量每行的元素数
29602961

2961-
LLAMA_LOG_DEBUG("222\n", 0);
29622962
for (int h = 0; h < n_head_kv; ++h) {
29632963
int k_start = h * (n_embd_head_qk_nope + n_embd_head_v);
29642964
for (int row = 0; row < kv_lora_rank; ++row) {
29652965
for (int col = 0; col < n_embd_head_qk_nope; ++col) {
2966-
LLAMA_LOG_DEBUG("333 row: %d, col: %d\n", row, col);
2966+
LLAMA_LOG_INFO("wk_b row: %d, col: %d\n", row, col);
29672967
int src_idx = row * src_stride + k_start + col;
29682968
GGML_ASSERT(src_idx < ggml_nelements(wkv_b));
29692969

29702970
int dst_row = h * kv_lora_rank + row;
29712971
int dst_col = col;
2972+
LLAMA_LOG_INFO("wk_b dst_row: %d, dst_col: %d\n", dst_row, dst_col);
29722973
dst[dst_row * n_embd_head_qk_nope + dst_col] = src[src_idx];
29732974
}
29742975
}
@@ -2981,25 +2982,25 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29812982
n_head_kv * n_embd_head_v, // 行数:合并头和特征维度
29822983
kv_lora_rank // 列数:LoRA 秩
29832984
);
2984-
LLAMA_LOG_DEBUG("444\n", 0);
2985+
LLAMA_LOG_INFO("wv_b shape: [%d, %d]\n", layer.wv_b->ne[0], layer.wv_b->ne[1]);
29852986
{
29862987
float *src = (float *)wkv_b->data;
29872988
float *dst = (float *)layer.wv_b->data;
29882989
int src_stride = wkv_b->ne[0]; // 原始张量每行的元素数
29892990

2990-
LLAMA_LOG_DEBUG("555\n", 0);
29912991
for (int h = 0; h < n_head_kv; ++h) {
29922992
int v_start = h * (n_embd_head_qk_nope + n_embd_head_v) + n_embd_head_qk_nope;
29932993
for (int row = 0; row < kv_lora_rank; ++row) {
29942994
for (int col = 0; col < n_embd_head_v; ++col) {
2995-
LLAMA_LOG_DEBUG("666 row: %d, col: %d\n", row, col);
2995+
LLAMA_LOG_INFO("wv_b row: %d, col: %d\n", row, col);
29962996
// 源索引计算
29972997
int src_idx = row * src_stride + v_start + col;
29982998
GGML_ASSERT(src_idx < ggml_nelements(wkv_b));
29992999

30003000
// 目标索引计算
30013001
int dst_row = h * n_embd_head_v + col; // 合并头和特征维度
30023002
int dst_col = row; // LoRA 秩维度
3003+
LLAMA_LOG_INFO("wv_b dst_row: %d, dst_col: %d\n", dst_row, dst_col);
30033004
dst[dst_row * kv_lora_rank + dst_col] = src[src_idx];
30043005
}
30053006
}

0 commit comments

Comments
 (0)