-
Notifications
You must be signed in to change notification settings - Fork 10.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
server : add "token healing" support #5765
Comments
The usual name for this feature is "token healing". I agree that it would be nice to have it supported here. |
@ggerganov I'd like to try working on it as my first issue! |
Ok. This can be demonstrated in one of the examples. One way would be to add it to |
@mare5x Sorry, I have not actually started so please don't wait for me. I'll try to take a look at your PR this week though and will be happy to help in any way I can. |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Feature Description
Hi! I am experimenting with using llama.cpp as a general-purpose code completion backend, similar to TabNine.
I am encountering a small problem: if the completion prompt ends mid-word, the results are not very accurate. For example, for a prompt such as
Five, Four, Thre
[sic], the model will often ignore the typo and suggest, Two
(formingThre, Two
).I think, as an option to the
/completion
server API, the following optional behavior would be useful:Thanks!
The text was updated successfully, but these errors were encountered: