Skip to content

Releases: CodeLinaro/llama.cpp

b5797

02 Jul 03:54
de56944
Compare
Choose a tag to compare
ci : disable fast-math for Metal GHA CI (#14478)

* ci : disable fast-math for Metal GHA CI

ggml-ci

* cont : remove -g flag

ggml-ci

b5752

24 Jun 19:03
62af464
Compare
Choose a tag to compare
batch : fix check for empty sequences in memory (#14364)

* batch : fix check for empty sequences in memory

ggml-ci

* cont : reuse the var

ggml-ci

b5689

18 Jun 05:56
c465030
Compare
Choose a tag to compare
cmake: remove shader-gen step-targets from ggml-vulkan (#14226)

* Remove step-targets from vulkan-shaders-gen

* Unset DESTDIR when building vulkan-shaders-gen

b5686

16 Jun 21:38
e434e69
Compare
Choose a tag to compare
common : suggest --jinja when autodetection fails (#14222)

b5627

10 Jun 20:37
3678b83
Compare
Choose a tag to compare
llama : support GEGLU for jina-bert-v2 (#14090)

b5548

30 May 23:37
e562eec
Compare
Choose a tag to compare
CUDA: fix typo in FlashAttention code (#13926)

b5460

23 May 00:11
3079e9a
Compare
Choose a tag to compare
release : fix windows hip release (#13707)

* release : fix windows hip release

* make single hip release with multiple targets

b5255

02 May 06:55
d24d592
Compare
Choose a tag to compare
ci: fix cross-compile sync issues (#12804)

b5098

10 Apr 21:36
64eda5d
Compare
Choose a tag to compare
convert : ability to lazy-load safetensors remotely without downloadi…

b5022

01 Apr 22:20
f423981
Compare
Choose a tag to compare
opencl : fix memory allocation size (#12649)

issue:
https://github.com/CodeLinaro/llama.cpp/pull/17#issuecomment-2760611283

This patch fixes the memory allocation size
not exceeding the maximum size of the OpenCL device.