Releases: ggerganov/llama.cpp
Releases · ggerganov/llama.cpp
b4273
vulkan: Add VK_NV_cooperative_matrix2 support for mul_mat and flash a…
b4272
llama : add Minerva 7B model support (#10673) * Support for Minerva 7B * Update convert_hf_to_gguf_update.py
b4271
sync : ggml
b4267
Update deprecation-warning.cpp (#10619) Fixed Path Separator Handling for Cross-Platform Support (Windows File Systems)
b4266
server : fix speculative decoding with context shift (#10641) * server : fix speculative decoding with context shift ggml-ci * server : take into account speculative limits ggml-ci * server : add tests
b4265
ggml : add predefined list of CPU backend variants to build (#10626) * ggml : add predefined list of CPU backend variants to build * update CPU dockerfiles
b4262
llama: Support MiniCPM-1B (with & w/o longrope) (#10559)
b4261
vulkan: Implement "fast divide" (mul+shift) for unary ops like copy (…
b4260
SYCL : Move to compile time oneMKL interface backend selection for NV…
b4258
Avoid using __fp16 on ARM with old nvcc (#10616)