Skip to content

Commit ab9a13a

Browse files
committed
fix
1 parent 70fb2f9 commit ab9a13a

File tree

1 file changed

+3
-5
lines changed

1 file changed

+3
-5
lines changed

src/llama-model.cpp

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2948,11 +2948,9 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
29482948

29492949
ggml_context * ctx = ctx_for_buft(buft);
29502950

2951-
auto trans_wkv_b = ggml_transpose(ctx, layer.wkv_b);
2952-
auto wkv_b_copied = ggml_new_tensor_2d(ctx, trans_wkv_b->type, trans_wkv_b->ne[0], trans_wkv_b->ne[1]);
2953-
ggml_cpy(ctx, trans_wkv_b, wkv_b_copied);
2954-
layer.wk_b = ggml_view_2d(ctx, wkv_b_copied, wkv_b_copied->ne[0], n_embd_head_qk_nope, n_head, 0);
2955-
layer.wv_b = ggml_view_2d(ctx, wkv_b_copied, wkv_b_copied->ne[0], n_embd_head_v, n_head, n_embd_head_qk_nope * n_head);
2951+
auto trans_wkv_b = ggml_cont(ctx, ggml_transpose(ctx, layer.wkv_b));
2952+
layer.wk_b = ggml_view_2d(ctx, trans_wkv_b, trans_wkv_b->ne[0], n_embd_head_qk_nope, n_head, 0);
2953+
layer.wv_b = ggml_view_2d(ctx, trans_wkv_b, trans_wkv_b->ne[0], n_embd_head_v, n_head, n_embd_head_qk_nope * n_head);
29562954
}
29572955
layer.wo = create_tensor(tn(LLM_TENSOR_ATTN_OUT, "weight", i), { n_head * ( n_embd_head_v), n_embd}, 0);
29582956

0 commit comments

Comments
 (0)