Bug: Coredump when quanting to Q4_0_*_* with imatrix #8767
Labels
bug-unconfirmed
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
What happened?
Hello,
I don't know if I am supposed to use imatrix with the new ARM dedicated quants (
Q4_0_*_*
). However, when I try to, I getAborted (core dumped)
.Is not using imatrix with those quants intentional? If that is the case, why does the quantization to
q4_*
andq5_*
work with imatrix?Name and Version
Rope scaling fix for L3.1 commit. Can't get the latest build due to a new build error I am investigating.
What operating system are you seeing the problem on?
Linux
Relevant log output
The text was updated successfully, but these errors were encountered: