We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Since the PR ggml-org/llama.cpp#9322 still has been settling, could you provide a Docker image for current version of https://github.com/OpenBMB/llama.cpp/tree/minicpm3?
I want to convert the model to .gguf format and quantize it, and it can be cool if there's no need for compiling the llama.cpp.
.gguf
llama.cpp
Or could you upload the quantized version in .gguf format to HuggingFace?
Thanks in advance!
The text was updated successfully, but these errors were encountered:
Hi, thanks for the attention. We have provided MiniCPM3 models in GGUF format (fp16 and q4). Check here.
Sorry, something went wrong.
No branches or pull requests
Feature request / 功能建议
Since the PR ggml-org/llama.cpp#9322 still has been settling, could you provide a Docker image for current version of https://github.com/OpenBMB/llama.cpp/tree/minicpm3?
I want to convert the model to
.gguf
format and quantize it, and it can be cool if there's no need for compiling thellama.cpp
.Or could you upload the quantized version in
.gguf
format to HuggingFace?Thanks in advance!
The text was updated successfully, but these errors were encountered: