Skip to content

Releases: ggerganov/llama.cpp

b4273

05 Dec 19:59
c9c6e01
Compare
Choose a tag to compare
vulkan: Add VK_NV_cooperative_matrix2 support for mul_mat and flash a…

b4272

05 Dec 19:25
6fe6247
Compare
Choose a tag to compare
llama : add Minerva 7B model support (#10673)

* Support for Minerva 7B

* Update convert_hf_to_gguf_update.py

b4271

05 Dec 12:12
0cd182e
Compare
Choose a tag to compare
sync : ggml

b4267

04 Dec 23:16
f112d19
Compare
Choose a tag to compare
Update deprecation-warning.cpp (#10619)

Fixed Path Separator Handling for Cross-Platform Support (Windows File Systems)

b4266

04 Dec 21:22
1da7b76
Compare
Choose a tag to compare
server : fix speculative decoding with context shift (#10641)

* server : fix speculative decoding with context shift

ggml-ci

* server : take into account speculative limits

ggml-ci

* server : add tests

b4265

04 Dec 14:49
59f4db1
Compare
Choose a tag to compare
ggml : add predefined list of CPU backend variants to build (#10626)

* ggml : add predefined list of CPU backend variants to build

* update CPU dockerfiles

b4262

04 Dec 10:25
8d0cfd5
Compare
Choose a tag to compare
llama: Support MiniCPM-1B (with & w/o longrope) (#10559)

b4261

04 Dec 08:02
2759916
Compare
Choose a tag to compare
vulkan: Implement "fast divide" (mul+shift) for unary ops like copy (…

b4260

04 Dec 02:15
40c6d79
Compare
Choose a tag to compare
SYCL : Move to compile time oneMKL interface backend selection for NV…

b4258

04 Dec 01:45
cd2f37b
Compare
Choose a tag to compare
Avoid using __fp16 on ARM with old nvcc (#10616)