Releases: CodeLinaro/llama.cpp
Releases · CodeLinaro/llama.cpp
b5460
b5255
ci: fix cross-compile sync issues (#12804)
b5098
convert : ability to lazy-load safetensors remotely without downloadi…
b5022
opencl : fix memory allocation size (#12649) issue: https://github.com/CodeLinaro/llama.cpp/pull/17#issuecomment-2760611283 This patch fixes the memory allocation size not exceeding the maximum size of the OpenCL device.
b4967
SYCL: implement memset ggml backend buffer interface (#12580) * SYCL: implement memset ggml backend buffer interface * use GGML_ABORT macro * Do not wait for all queues to finish for memset operation
b4951
opencl: simplify kernel embedding logic in cmakefile (#12503) Co-authored-by: Max Krasnyansky <[email protected]>
b4903
vulkan: Add N/2 and N/4 optimized paths in coopmat2 shader (#12312)
b4719
llguidance build fixes for Windows (#11664) * setup windows linking for llguidance; thanks @phil-scott-78 * add build instructions for windows and update script link * change VS Community link from DE to EN * whitespace fix
b4717
cuda : add ampere to the list of default architectures (#11870)
b4667
Make logging more verbose (#11714) Debugged an issue with a user who was on a read-only filesystem. Signed-off-by: Eric Curtin <[email protected]>