Releases: unslothai/llama.cpp
Releases · unslothai/llama.cpp
b5888
Merge branch 'ggml-org:master' into master
b5885
Fixes
b5884
tests : cover lfm2 cases in test_ssm_conv (#14651)
b5873
model : support LiquidAI LFM2 hybrid family (#14620) **Important** LFM2 was [merged ](https://github.com/huggingface/transformers/pull/39340)into transformers, but has not yet been released. To convert into gguf, install transformers from source ```shell pip install "transformers @ git+https://github.com/huggingface/transformers.git@main" ```
b5334
Merge branch 'ggml-org:master' into master
b5319
Update getrows.cu
b5318
ci : limit write permission to only the release step + fixes (#13392) * ci : limit write permission to only the release step * fix win cuda file name * fix license file copy on multi-config generators
b5287
CUDA: fix logic for clearing padding with -ngl 0 (#13320)
b5272
imatrix: fix oob writes if src1 is not contiguous (#13286)
b5270
Revert "CUDA: batched+noncont MMQ, refactor bs>1 MoE code (#13199)" This reverts commit e1e8e0991ffd9e99a445c6812bb519d5bac9f4b5.