-
Notifications
You must be signed in to change notification settings - Fork 51
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ramalama does not work with granite models #338
Comments
Fix: This is specifically a problem with granitemoe models |
FYI not just
|
It's likely the same fix, update llama.cpp , it's amazing how quickly upstream llama.cpp project moves, if you add this patch at least for the granite-moe case you get this:
because this model architecture was added recently
|
This should be fixed with release v0.0.20 |
Ollama announced support for IBM Granite https://x.com/ollama/status/1848223852465213703
I tried to run
granite3-moe
with ramalamaBut it fails after download without printing any error log. Latest Ollama works fine with this model
OS: Ubuntu 23.10
Python: 3.11.6
Ramalama:
The text was updated successfully, but these errors were encountered: