Skip to content

Commit 7edf380

Browse files
Merge branch 'add-fh1-rebased' of https://github.com/tiiuae/llama.cpp-public into add-fh1-rebased
2 parents 9b92648 + 90ddf24 commit 7edf380

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

src/llama-arch.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -129,7 +129,7 @@ static const std::map<llm_kv, const char *> LLM_KV_NAMES = {
129129
{ LLM_KV_EMBEDDING_SCALE, "%s.embedding_scale" },
130130
{ LLM_KV_TOKEN_SHIFT_COUNT, "%s.token_shift_count" },
131131
{ LLM_KV_INTERLEAVE_MOE_LAYER_STEP, "%s.interleave_moe_layer_step" },
132-
132+
133133
{ LLM_KV_ATTENTION_HEAD_COUNT, "%s.attention.head_count" },
134134
{ LLM_KV_ATTENTION_HEAD_COUNT_KV, "%s.attention.head_count_kv" },
135135
{ LLM_KV_ATTENTION_MAX_ALIBI_BIAS, "%s.attention.max_alibi_bias" },

src/llama-hparams.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -74,8 +74,8 @@ uint32_t llama_hparams::n_embd_r() const {
7474
// TODO: maybe support other convolution strides than 1
7575
// NOTE: since the first column of the conv_state is shifted out each time, it's not actually needed
7676
// Corresponds to Mamba's conv_states size
77-
78-
// check if the architecture is using d_ssm
77+
78+
// check if the architecture is using d_ssm
7979
return (ssm_d_conv > 0 ? ssm_d_conv - 1 : 0) * (ssm_d_inner + 2*ssm_n_group*ssm_d_state);
8080
}
8181

src/llama-model.cpp

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -4600,7 +4600,7 @@ bool llama_model::load_tensors(llama_model_loader & ml) {
46004600
if (output == NULL) {
46014601
output = create_tensor(tn(LLM_TENSOR_TOKEN_EMBD, "weight"), {n_embd, n_vocab}, TENSOR_DUPLICATED);
46024602
}
4603-
4603+
46044604
for (int i = 0; i < n_layer; ++i) {
46054605
auto & layer = layers[i];
46064606
layer.attn_norm = create_tensor(tn(LLM_TENSOR_ATTN_NORM, "weight", i), {n_embd}, 0);
@@ -14738,7 +14738,7 @@ struct llm_build_falcon_h1 : public llm_graph_context {
1473814738
inpSA = ggml_add(ctx0, cur, inpSA);
1473914739
cb(cur, "layer_out", il);
1474014740

14741-
if (il == n_layer - 1 && inp_out_ids) {
14741+
if (il == n_layer - 1 && inp_out_ids) {
1474214742
cur = ggml_get_rows(ctx0, cur, inp_out_ids);
1474314743
inpSA = ggml_get_rows(ctx0, inpSA, inp_out_ids);
1474414744
}

0 commit comments

Comments
 (0)