-
Notifications
You must be signed in to change notification settings - Fork 11k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Deepseek-based model throws std::out_of_range
exception on load
#5688
Comments
whoops, it was indeed my mistake in the conversion! turns out that, while the base instruct model uses a fast tokenizer, this model instead uses the regular llama tokenizer. which means I should've converted with BPE! reconverted & quantized and what do you know, it runs great. I doubt it's worth investigating the crash on its own given the incorrectly produced model file. But maybe there could be a way to detect this sort of mistake at conversion time to short circuit this process? Auto vocab-type detection would be beneficial, but that's out of the scope of this issue. |
Huh, that's surprising. There is a long pending PR that I thought needs to be merged to support DeepSeek models: #5464. It should fix some tokenization problems AFAICT + add conversion I'm surprised that it worked for you |
the updated fp16 conversion and quants just finished uploading: hf link it does seem to work fine tho! I haven't tested it too extensively, but: full output running main:
that script just calls main with in-prefix/-suffix, ngl, temp, etc. |
Model: OpenCodeInterpreter-DS-6.7B (GGUFs)
This is a deepseek coder instruct-based model, llama arch, but maybe there's something distinct for it that requires special-handling?
Or maybe I did something wrong in converting these files from the original safetensors (used the same build, b2249, for converting, quantizing, and running).
Both
-ngl=999
&-ngl=0
produce the same exception:llama.cpp build info
b2249
(rev:15499eb94227401bdc8875da6eb85c15d37068f7
)LLAMA_METAL=1
lldb stacktrace
full lldb output from `./main`:
conversion info
The text was updated successfully, but these errors were encountered: