Skip to content

Commit 0c92f36

Browse files
committed
add llama i
add llama
1 parent ff545df commit 0c92f36

File tree

7 files changed

+859
-31
lines changed

7 files changed

+859
-31
lines changed

open_diloco/configs/config_1b.json

Lines changed: 6 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,9 @@
11
{
2-
"architectures": [
3-
"LlamaForCausalLM"
4-
],
5-
"model_type": "llama",
6-
"hidden_size": 2048,
2+
"name": "llama",
3+
"n_embd": 2048,
74
"intermediate_size": 5632,
8-
"num_attention_heads": 32,
9-
"num_hidden_layers": 22,
10-
"use_cache": false,
11-
"rms_norm_eps": 1e-05,
12-
"num_key_value_heads": 4
5+
"n_head": 32,
6+
"n_layer": 22,
7+
"n_query_groups": 4,
8+
"vocab_size": 1024
139
}

open_diloco/configs/config_2m.json

Lines changed: 5 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,10 @@
11
{
2-
"architectures": [
3-
"LlamaForCausalLM"
4-
],
5-
"model_type": "llama",
6-
"hidden_size": 64,
2+
"name": "llama_2m",
3+
"n_embd": 64,
74
"intermediate_size": 256,
8-
"num_attention_heads": 2,
9-
"num_hidden_layers": 2,
10-
"rms_norm_eps": 1e-05,
11-
"use_cache": false,
5+
"n_head": 2,
6+
"n_layer": 2,
127
"vocab_size": 1024
13-
}
8+
}
149

1510

0 commit comments

Comments
 (0)