Skip to content

Commit b092a2c

Browse files
committed
tmp
1 parent fd32a43 commit b092a2c

File tree

1 file changed

+3
-2
lines changed

1 file changed

+3
-2
lines changed

src/llama-model.cpp

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2949,8 +2949,9 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29492949
ggml_context * ctx = ctx_for_buft(buft);
29502950

29512951
auto trans_wkv_b = ggml_transpose(ctx, layer.wkv_b);
2952-
layer.wk_b = ggml_view_2d(ctx, trans_wkv_b, trans_wkv_b->ne[0], n_embd_head_qk_nope, n_head, 0);
2953-
layer.wv_b = ggml_view_2d(ctx, trans_wkv_b, trans_wkv_b->ne[0], n_embd_head_v, n_head, n_embd_head_qk_nope * n_head);
2952+
auto wkv_b_copied = ggml_dup(ctx, trans_wkv_b);
2953+
layer.wk_b = ggml_view_2d(ctx, wkv_b_copied, wkv_b_copied->ne[0], n_embd_head_qk_nope, n_head, 0);
2954+
layer.wv_b = ggml_view_2d(ctx, wkv_b_copied, wkv_b_copied->ne[0], n_embd_head_v, n_head, n_embd_head_qk_nope * n_head);
29542955
}
29552956
layer.wo = create_tensor(tn(LLM_TENSOR_ATTN_OUT, "weight", i), { n_head * ( n_embd_head_v), n_embd}, 0);
29562957

0 commit comments

Comments
 (0)