-
Notifications
You must be signed in to change notification settings - Fork 10.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Feature Request: Better chat UX for llama-cli #11202
Comments
I think it's OK to repurpose #11203 looks OK to me. IMO we can make a quick pass to strip |
I disagree about breaking |
Prefix and suffix are definitely needed in As for antiprompts, aren't they necessary for base (not instruct) models and custom instructions? Would |
Yes, I guess the functions such as |
The main problem with My plan is trying not to breaking many things while doing #11203. @MaggotHATE Suffix, prefix are easy to add back, I totally understand that many users still use them for alpaca-style chatbots. Maybe other things that need to worry about too, but I'll make a list once we're more clear about the global direction.
Yup, that's one way. But the problem is that most guides that I can find on internet already told users to use A hypothetical better plan could be:
|
To be honest, I'm not sure it's a good idea to add I would rather suggest starting from slight refurbishment of |
What I mean by using
And no, I'm not encouraging developers to use it in downstream projects. My point is that
My POV is that it's much harder than saying. The (For example, the |
Is it bad though? I'm sure it's a valid way to organize inference in a chat app, but it is definitely less convenient to work with. On the other hand, it's more transparent in its own way, so maybe there's a value in keeping I need to look at the most recent version again and see what can be done. |
Feature Description
llama-cli
in chat (conversation) mode automatically if there is a built-in chat template/regen
,/readfile
, etc (demo in main : add new feature: special commands #10145)Motivation
Inspired by my first attempt at adding commands to
llama-cli
, it seems thatmain.cpp
is a bit too complex for a chat-based application.Now, you might ask: don’t we already have
llama-run
andsimple-chat
for that? True, but the issue is that these examples are largely standalone - they don’t rely oncommon.cpp
.Given that most users are already familiar with
llama-cli
, enhancing the chat UX within it could benefit a majority of users, including those who don’t actively follow llama.cpp’s development. This approach is similar to what we did with the updated UI forllama-server
: a seamless, drop-in replacement that eliminates the need for users to change their workflows.Possible Implementation
One way to do is to split all the chat-related logic into
main/chat.hpp
, then call it frommain.cpp
, see #11203CC @ggerganov and @slaren for discussion, thanks!
The text was updated successfully, but these errors were encountered: