Releases: CodeLinaro/llama.cpp
Releases · CodeLinaro/llama.cpp
b5797
b5752
batch : fix check for empty sequences in memory (#14364) * batch : fix check for empty sequences in memory ggml-ci * cont : reuse the var ggml-ci
b5689
cmake: remove shader-gen step-targets from ggml-vulkan (#14226) * Remove step-targets from vulkan-shaders-gen * Unset DESTDIR when building vulkan-shaders-gen
b5686
common : suggest --jinja when autodetection fails (#14222)
b5627
llama : support GEGLU for jina-bert-v2 (#14090)
b5548
CUDA: fix typo in FlashAttention code (#13926)
b5460
release : fix windows hip release (#13707) * release : fix windows hip release * make single hip release with multiple targets
b5255
ci: fix cross-compile sync issues (#12804)
b5098
convert : ability to lazy-load safetensors remotely without downloadi…
b5022
opencl : fix memory allocation size (#12649) issue: https://github.com/CodeLinaro/llama.cpp/pull/17#issuecomment-2760611283 This patch fixes the memory allocation size not exceeding the maximum size of the OpenCL device.