Skip to content

Commit 11b0b54

Browse files
fix(llama.cu): 改正依赖更新导致的编译问题
Signed-off-by: PanZezhong <[email protected]>
1 parent 137c398 commit 11b0b54

File tree

3 files changed

+16
-12
lines changed

3 files changed

+16
-12
lines changed

Cargo.lock

Lines changed: 12 additions & 11 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -129,7 +129,8 @@ Options:
129129
path = "model-path"
130130
gpus = [0, 1, 2, 3]
131131
max-tokens = 4096
132-
think = true
133132
temperature = 0.9
134133
top-p = 0.6
134+
repetition_penalty = 1.02
135+
think = true
135136
```

llama.cu/src/model/qw2vl_mmproj.rs

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -52,6 +52,8 @@ impl GGufModel<'_> {
5252
get(&format!("v.blk.{iblk}.attn_qkv.weight")),
5353
Some((dt_norm, get(&format!("v.blk.{iblk}.attn_qkv.bias")))),
5454
),
55+
q_norm: None,
56+
k_norm: None,
5557
rope: Some(RoPE {
5658
multimodal: true,
5759
nctx: 34, // todo: from image

0 commit comments

Comments
 (0)