-
Notifications
You must be signed in to change notification settings - Fork 40
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Q] How do I use a base (completion only, no instruct/chat) model? #108
Comments
Indeed, looking at the logs, ellama is using a template:
|
ellama seems to be parsing the response as well and doing magic on it. (E.g., the response gets suddenly deleted when it finishes streaming, presumably because it wasn't in backticks.) I want to disable all magic. |
You can use |
@s-kostyaev Thanks. Is there a way to limit the completion so that it stops on a newline? |
@NightMachinery Sure. You need to create custom model with ollama. Add parameter:
And create custom model from this modelfile. And use this new created model. For example, I use https://ollama.com/sskostyaev/openchat:1l to create chat names. |
@s-kostyaev Looking at the logs,
This completion API works pretty good for completing text in my tests. This API with a reasonable |
I have set:
But ellama seems to be sending some kind of template-y message to ollama when I do
ellama-code-complete
. I don't want any "prompt engineering", I just want to feed the context near point into this base model and get its next N lines of prediction.The text was updated successfully, but these errors were encountered: