-
Notifications
You must be signed in to change notification settings - Fork 11k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Gemma models quantized using llamacpp not working in lm studio #5706
Comments
I can only speak for myself but I 100% refuse to debug a problem unless it can be reproduced entirely with open-source code. |
No it literally doesnt work I just built this version of llamacpp. And that .py script doesnt work for Gemma Plus that not the script you are even suppose to use according to the documentation. You are suppose to use convert.py
|
@JohannesGaessler I totally understand that, hopefully openai is willing to reach out and work with you to fix this |
Im uploading the model files for the merges if anyone wants to do some debugging. Should be in the next 10 hours or so. Sorry slow internet. Follow the mulit-thread. And check out my model for debugging. Thread links: |
@rombodawg did you try latest version? it already support |
This issue was closed because it has been inactive for 14 days since being marked as stale. |
I think like me people are considering other options. I will however keep on watching the release notes to see when this is fixed. |
Gemma models that have been quantized using Llamacpp are not working. Please look into the issue
error
"llama.cpp error: 'create_tensor: tensor 'output.weight' not found'"
I will open a issue on the lm studio github aswell addressing this
lmstudio-ai/configs#21
System:
Ryzen 5600x
rtx 3080 gpu
b550 motherboard
64gb ddr4 ram
windows 10 OS
The text was updated successfully, but these errors were encountered: