Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why can't multiple apis be triggered at the same time #873

Open
zhengzhanpeng opened this issue Nov 5, 2023 · 4 comments
Open

Why can't multiple apis be triggered at the same time #873

zhengzhanpeng opened this issue Nov 5, 2023 · 4 comments
Labels
question Further information is requested

Comments

@zhengzhanpeng
Copy link

This is an area that has been a headache for me, and I hope someone can answer me

@abetlen abetlen added the question Further information is requested label Nov 8, 2023
@tranhoangnguyen03
Copy link

Do you mean concurrent requests?

@zpzheng
Copy link

zpzheng commented Nov 17, 2023

Do you mean concurrent requests?

yes

@abetlen
Copy link
Owner

abetlen commented Nov 23, 2023

@zhengzhanpeng this currently under development in #771

@isaac-chung
Copy link

@abetlen I noticed that llama.cpp's server supports concurrent requests and continuous batching as well https://github.com/ggerganov/llama.cpp/tree/master/examples/server. To enable that for this library, would it be as straightforward as exposing the relevant command line options? Or am I missing something obvious?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

5 participants