Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

not enough space in the buffer (needed 442368, largest block available 290848) GGML_ASSERT: #33

Open
adi1999 opened this issue Jan 19, 2024 · 3 comments

Comments

@adi1999
Copy link

adi1999 commented Jan 19, 2024

Just for those facing this issue while trying to run a local LLM - revert llama-cpp-python to 0.2.0 , pip install llama-cpp-python==0.2.0

@fynnfluegge
Copy link
Owner

Hey @adi1999, good catch! Would it makes sense to add the explicit version to the installation command in llm.py? Do you know what is the cause of this issue?

@adi1999
Copy link
Author

adi1999 commented Jan 24, 2024

Hey @fynnfluegge , maybe coz the model that we're using is trained on bigger context size, while llama.cpp sets 512 by default, so either we can increase it manually, or using this specific version is a workaround.
Btw, you've done awesome work :)

@fynnfluegge
Copy link
Owner

Alright, thank you @adi1999 !

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants