@@ -6172,9 +6172,9 @@ static bool llm_load_tensors(
61726172 layer.ffn_up = ml.create_tensor(ctx_split, tn(LLM_TENSOR_FFN_UP, "weight", i), {n_embd, n_ff});
61736173
61746174 // optional MLP bias
6175- layer.ffn_gate_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_GATE, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6176- layer.ffn_down_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_DOWN, "bias", i), {n_embd}, llama_model_loader::TENSOR_NOT_REQUIRED);
6177- layer.ffn_up_b = ml.create_tensor(ctx_split , tn(LLM_TENSOR_FFN_UP, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6175+ layer.ffn_gate_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_GATE, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
6176+ layer.ffn_down_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_DOWN, "bias", i), {n_embd}, llama_model_loader::TENSOR_NOT_REQUIRED);
6177+ layer.ffn_up_b = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_FFN_UP, "bias", i), {n_ff}, llama_model_loader::TENSOR_NOT_REQUIRED);
61786178 } else {
61796179 layer.ffn_gate_inp = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_FFN_GATE_INP, "weight", i), {n_embd, n_expert});
61806180
@@ -6498,7 +6498,7 @@ static bool llm_load_tensors(
64986498 layer.bv = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_V, "bias", i), {n_embd_gqa});
64996499
65006500 layer.wo = ml.create_tensor(ctx_split, tn(LLM_TENSOR_ATTN_OUT, "weight", i), {n_embd, n_embd}); //output_dens
6501- layer.bo = ml.create_tensor(ctx_split , tn(LLM_TENSOR_ATTN_OUT, "bias", i), {n_embd}); //output_dens
6501+ layer.bo = ml.create_tensor(ctx_layer , tn(LLM_TENSOR_ATTN_OUT, "bias", i), {n_embd}); //output_dens
65026502
65036503 layer.attn_out_norm = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_OUT_NORM, "weight", i), {n_embd}); //output_norm
65046504 layer.attn_out_norm_b = ml.create_tensor(ctx_layer, tn(LLM_TENSOR_ATTN_OUT_NORM, "bias", i), {n_embd});
0 commit comments