Skip to content

Releases: unslothai/llama.cpp

b5888

13 Jul 06:26
a032396

Choose a tag to compare

Merge branch 'ggml-org:master' into master

b5885

13 Jul 06:19

Choose a tag to compare

Fixes

b5884

13 Jul 02:20
c31e606

Choose a tag to compare

tests : cover lfm2 cases in test_ssm_conv (#14651)

b5873

12 Jul 08:16
f5e96b3

Choose a tag to compare

model : support LiquidAI LFM2 hybrid family (#14620)

**Important**
LFM2 was [merged ](https://github.com/huggingface/transformers/pull/39340)into transformers, but has not yet been released.
To convert into gguf, install transformers from source
```shell
pip install "transformers @ git+https://github.com/huggingface/transformers.git@main"
```

b5334

09 May 22:57
0d90bbe

Choose a tag to compare

Merge branch 'ggml-org:master' into master

b5319

08 May 23:38
f688555

Choose a tag to compare

Update getrows.cu

b5318

08 May 23:24
15e0328

Choose a tag to compare

ci : limit write permission to only the release step + fixes (#13392)

* ci : limit write permission to only the release step

* fix win cuda file name

* fix license file copy on multi-config generators

b5287

06 May 00:06
9070365

Choose a tag to compare

CUDA: fix logic for clearing padding with -ngl 0 (#13320)

b5272

04 May 01:20
3e959f0

Choose a tag to compare

imatrix: fix oob writes if src1 is not contiguous (#13286)

b5270

03 May 08:24

Choose a tag to compare

Revert "CUDA: batched+noncont MMQ, refactor bs>1 MoE code (#13199)"

This reverts commit e1e8e0991ffd9e99a445c6812bb519d5bac9f4b5.