Commit 73cf442
llama : fix Gemma-2 Query scaling factors (ggml-org#8473)
* 9B - query_pre_attn_scalar = 256 not 224
See google/gemma_pytorch@03e6575
Gemma 9b should use 256 and not 224 (self.config.hidden_size // self.config.num_attention_heads)
* llama : fix Gemma-2 Query scaling factor
ggml-ci
---------
Co-authored-by: Daniel Han <[email protected]>1 parent e236528 commit 73cf442
2 files changed
+6
-6
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
2504 | 2504 | | |
2505 | 2505 | | |
2506 | 2506 | | |
2507 | | - | |
2508 | | - | |
2509 | | - | |
2510 | | - | |
2511 | | - | |
2512 | 2507 | | |
2513 | 2508 | | |
2514 | 2509 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
11680 | 11680 | | |
11681 | 11681 | | |
11682 | 11682 | | |
11683 | | - | |
| 11683 | + | |
| 11684 | + | |
| 11685 | + | |
| 11686 | + | |
| 11687 | + | |
| 11688 | + | |
11684 | 11689 | | |
11685 | 11690 | | |
11686 | 11691 | | |
| |||
0 commit comments