Skip to content

Releases: ggerganov/llama.cpp

b4659

07 Feb 08:11
225bbbf
Compare
Choose a tag to compare
ggml : optimize and build warning fix for LoongArch (#11709)

* ggml : optimize convert f32<->f16 for loongarch_asx

* ggml : optimize loongarch_asx extend i16,i8,u8 to i32,i16

* ggml : Fix warnings when run cpu CI locally on LoongArch

b4658

06 Feb 22:22
855cd07
Compare
Choose a tag to compare
llama : fix old glm4 models (#11670)

b4657

06 Feb 20:13
8a59053
Compare
Choose a tag to compare
sync : ggml

b4651

06 Feb 12:09
c0d4843
Compare
Choose a tag to compare
build : fix llama.pc (#11658)

Signed-off-by: Adrien Gallouët <[email protected]>

b4649

06 Feb 06:54
2c6c8df
Compare
Choose a tag to compare
vulkan: optimize coopmat2 iq2/iq3 callbacks (#11521)

* vulkan: optimize coopmat2 iq2/iq3 callbacks

* build: trigger CI on GLSL compute shader changes

b4648

06 Feb 06:55
8a7e3bf
Compare
Choose a tag to compare
vulkan: initial support for IQ4_XS quantization (#11501)

b4647

06 Feb 06:36
1b598b3
Compare
Choose a tag to compare
vulkan: use smaller combined allocations to avoid fragmentation (#11551)

b4646

06 Feb 02:22
902368a
Compare
Choose a tag to compare
metal : avoid breaking build when metal API predates TARGET_OS_VISION…

b4644

05 Feb 09:37
d774ab3
Compare
Choose a tag to compare
metal : adjust support conditions for norm operators (#11671)

cont #11659

ggml-ci

b4643

05 Feb 08:59
fa62da9
Compare
Choose a tag to compare
CUDA: support for mat. mul. with ne03 != ne13 (#11656)