@@ -2946,29 +2946,30 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29462946 buft = ggml_backend_dev_buffer_type (cpu_dev);
29472947 }
29482948
2949+ LLAMA_LOG_INFO (" n_head_kv: %d, kv_lora_rank: %d, n_embd_head_qk_nope: %d\n " , n_head_kv, kv_lora_rank, n_embd_head_qk_nope);
29492950 ggml_context * ctx = ctx_for_buft (buft);
29502951 layer.wk_b = ggml_new_tensor_2d (ctx,
29512952 wkv_b->type ,
29522953 n_head_kv * kv_lora_rank,
29532954 n_embd_head_qk_nope
29542955 );
2955- LLAMA_LOG_DEBUG ( " 111 \n " , 0 );
2956+ LLAMA_LOG_INFO ( " wk_b shape: [%d, %d] \n " , layer. wk_b -> ne [ 0 ], layer. wk_b -> ne [ 1 ] );
29562957 {
29572958 float *src = (float *)wkv_b->data ;
29582959 float *dst = (float *)layer.wk_b ->data ;
29592960 int src_stride = wkv_b->ne [0 ]; // 原始张量每行的元素数
29602961
2961- LLAMA_LOG_DEBUG (" 222\n " , 0 );
29622962 for (int h = 0 ; h < n_head_kv; ++h) {
29632963 int k_start = h * (n_embd_head_qk_nope + n_embd_head_v);
29642964 for (int row = 0 ; row < kv_lora_rank; ++row) {
29652965 for (int col = 0 ; col < n_embd_head_qk_nope; ++col) {
2966- LLAMA_LOG_DEBUG ( " 333 row: %d, col: %d\n " , row, col);
2966+ LLAMA_LOG_INFO ( " wk_b row: %d, col: %d\n " , row, col);
29672967 int src_idx = row * src_stride + k_start + col;
29682968 GGML_ASSERT (src_idx < ggml_nelements (wkv_b));
29692969
29702970 int dst_row = h * kv_lora_rank + row;
29712971 int dst_col = col;
2972+ LLAMA_LOG_INFO (" wk_b dst_row: %d, dst_col: %d\n " , dst_row, dst_col);
29722973 dst[dst_row * n_embd_head_qk_nope + dst_col] = src[src_idx];
29732974 }
29742975 }
@@ -2981,25 +2982,25 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29812982 n_head_kv * n_embd_head_v, // 行数:合并头和特征维度
29822983 kv_lora_rank // 列数:LoRA 秩
29832984 );
2984- LLAMA_LOG_DEBUG ( " 444 \n " , 0 );
2985+ LLAMA_LOG_INFO ( " wv_b shape: [%d, %d] \n " , layer. wv_b -> ne [ 0 ], layer. wv_b -> ne [ 1 ] );
29852986 {
29862987 float *src = (float *)wkv_b->data ;
29872988 float *dst = (float *)layer.wv_b ->data ;
29882989 int src_stride = wkv_b->ne [0 ]; // 原始张量每行的元素数
29892990
2990- LLAMA_LOG_DEBUG (" 555\n " , 0 );
29912991 for (int h = 0 ; h < n_head_kv; ++h) {
29922992 int v_start = h * (n_embd_head_qk_nope + n_embd_head_v) + n_embd_head_qk_nope;
29932993 for (int row = 0 ; row < kv_lora_rank; ++row) {
29942994 for (int col = 0 ; col < n_embd_head_v; ++col) {
2995- LLAMA_LOG_DEBUG ( " 666 row: %d, col: %d\n " , row, col);
2995+ LLAMA_LOG_INFO ( " wv_b row: %d, col: %d\n " , row, col);
29962996 // 源索引计算
29972997 int src_idx = row * src_stride + v_start + col;
29982998 GGML_ASSERT (src_idx < ggml_nelements (wkv_b));
29992999
30003000 // 目标索引计算
30013001 int dst_row = h * n_embd_head_v + col; // 合并头和特征维度
30023002 int dst_col = row; // LoRA 秩维度
3003+ LLAMA_LOG_INFO (" wv_b dst_row: %d, dst_col: %d\n " , dst_row, dst_col);
30033004 dst[dst_row * kv_lora_rank + dst_col] = src[src_idx];
30043005 }
30053006 }
0 commit comments