@@ -7757,9 +7757,9 @@ static bool llm_load_tensors(
77577757 }
77587758 } break;
77597759 case LLM_ARCH_DECI:
7760- {
7760+ {
77617761 model.tok_embd = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, 0);
7762-
7762+
77637763 // output
77647764 model.output_norm = create_tensor(tn(LLM_TENSOR_OUTPUT_NORM, "weight"), {n_embd}, 0);
77657765 model.output = create_tensor(tn(LLM_TENSOR_OUTPUT, "weight"), {n_embd, n_vocab}, llama_model_loader::TENSOR_NOT_REQUIRED);
@@ -11028,9 +11028,9 @@ struct llm_build_context {
1102811028 }
1102911029
1103011030 struct ggml_cgraph * build_deci() {
11031- struct ggml_cgraph * gf = ggml_new_graph_custom(ctx0, llama_model_max_nodes(model), false);
11031+ struct ggml_cgraph * gf = ggml_new_graph_custom(ctx0, llama_model_max_nodes(model), false);
1103211032
11033- // mutable variable, needed during the last layer of the computation to skip unused tokens
11033+ // mutable variable, needed during the last layer of the computation to skip unused tokens
1103411034 int32_t n_tokens = this->n_tokens;
1103511035
1103611036 const int64_t n_embd_head = hparams.n_embd_head_v;
@@ -11054,9 +11054,9 @@ struct llm_build_context {
1105411054 const int64_t n_head_kv = hparams.n_head_kv(il);
1105511055 const int64_t n_head = hparams.n_head(il);
1105611056
11057- if (n_head == 0) // attention-free layer of Llama-3_1-Nemotron-51B
11057+ if (n_head == 0) { // attention-free layer of Llama-3_1-Nemotron-51B
1105811058 cur = inpL;
11059- else {
11059+ } else {
1106011060 // norm
1106111061 cur = llm_build_norm(ctx0, inpL, hparams,
1106211062 model.layers[il].attn_norm, NULL,
@@ -11067,9 +11067,8 @@ struct llm_build_context {
1106711067 if (n_head > 0 && n_head_kv == 0) { // "linear attention" of Llama-3_1-Nemotron-51B
1106811068 cur = llm_build_lora_mm(lctx, ctx0, model.layers[il].wo, cur);
1106911069 cb(cur, "wo", il);
11070- } else if (n_head > 0)
11070+ } else if (n_head > 0) {
1107111071 // self-attention
11072- {
1107311072 // rope freq factors for llama3; may return nullptr for llama2 and other models
1107411073 struct ggml_tensor * rope_factors = build_rope_factors(il);
1107511074
0 commit comments