Commit 1dc9614
llama : fix kq_scale for the attention layers of PLaMo2 (ggml-org#14892)
* Fix dimensions for expand
* Change dimensions to copy states to cache
* Fix the default value for plamo2 conversion
* Fix scale given to build_attn
* Update src/llama-model.cpp
Co-authored-by: Sigbjørn Skjæret <[email protected]>
* Update src/llama-model.cpp
Co-authored-by: Sigbjørn Skjæret <[email protected]>
* Update src/llama-model.cpp
Co-authored-by: Sigbjørn Skjæret <[email protected]>
---------
Co-authored-by: Sigbjørn Skjæret <[email protected]>1 parent 446595b commit 1dc9614
2 files changed
+10
-9
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
3791 | 3791 | | |
3792 | 3792 | | |
3793 | 3793 | | |
3794 | | - | |
| 3794 | + | |
3795 | 3795 | | |
3796 | 3796 | | |
3797 | 3797 | | |
| |||
3802 | 3802 | | |
3803 | 3803 | | |
3804 | 3804 | | |
3805 | | - | |
| 3805 | + | |
3806 | 3806 | | |
3807 | 3807 | | |
3808 | 3808 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
16191 | 16191 | | |
16192 | 16192 | | |
16193 | 16193 | | |
16194 | | - | |
| 16194 | + | |
16195 | 16195 | | |
16196 | 16196 | | |
16197 | 16197 | | |
| |||
16231 | 16231 | | |
16232 | 16232 | | |
16233 | 16233 | | |
16234 | | - | |
| 16234 | + | |
16235 | 16235 | | |
16236 | 16236 | | |
16237 | 16237 | | |
| |||
16306 | 16306 | | |
16307 | 16307 | | |
16308 | 16308 | | |
16309 | | - | |
16310 | | - | |
| 16309 | + | |
| 16310 | + | |
| 16311 | + | |
16311 | 16312 | | |
16312 | 16313 | | |
16313 | 16314 | | |
| |||
16370 | 16371 | | |
16371 | 16372 | | |
16372 | 16373 | | |
16373 | | - | |
16374 | | - | |
16375 | | - | |
| 16374 | + | |
| 16375 | + | |
| 16376 | + | |
16376 | 16377 | | |
16377 | 16378 | | |
16378 | 16379 | | |
| |||
0 commit comments