Skip to content

Releases: unslothai/llama.cpp

b5888

13 Jul 06:26
a032396
Compare
Choose a tag to compare
Merge branch 'ggml-org:master' into master

b5885

13 Jul 06:19
Compare
Choose a tag to compare
Fixes

b5884

13 Jul 02:20
c31e606
Compare
Choose a tag to compare
tests : cover lfm2 cases in test_ssm_conv (#14651)

b5873

12 Jul 08:16
f5e96b3
Compare
Choose a tag to compare
model : support LiquidAI LFM2 hybrid family (#14620)

**Important**
LFM2 was [merged ](https://github.com/huggingface/transformers/pull/39340)into transformers, but has not yet been released.
To convert into gguf, install transformers from source
```shell
pip install "transformers @ git+https://github.com/huggingface/transformers.git@main"
```

b5334

09 May 22:57
0d90bbe
Compare
Choose a tag to compare
Merge branch 'ggml-org:master' into master

b5319

08 May 23:38
f688555
Compare
Choose a tag to compare
Update getrows.cu

b5318

08 May 23:24
15e0328
Compare
Choose a tag to compare
ci : limit write permission to only the release step + fixes (#13392)

* ci : limit write permission to only the release step

* fix win cuda file name

* fix license file copy on multi-config generators

b5287

06 May 00:06
9070365
Compare
Choose a tag to compare
CUDA: fix logic for clearing padding with -ngl 0 (#13320)

b5272

04 May 01:20
3e959f0
Compare
Choose a tag to compare
imatrix: fix oob writes if src1 is not contiguous (#13286)

b5270

03 May 08:24
Compare
Choose a tag to compare
Revert "CUDA: batched+noncont MMQ, refactor bs>1 MoE code (#13199)"

This reverts commit e1e8e0991ffd9e99a445c6812bb519d5bac9f4b5.