-
Notifications
You must be signed in to change notification settings - Fork 49
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
What other library that optimum-benchmark support other than transformer #117
Comments
I want to evaluate the inference latency, throughput, parameter numbers of a custom llm |
|
@L1-M1ng I would love to review a PR with llama.cpp support, https://github.com/abetlen/llama-cpp-python seems to be the most starred python bindings |
llama cpp support added in #231 🚀 |
@IlyasMoutawwakil , I am trying to run
|
Can I use Optimum-benchmark to evaluate the performance of qwen.cpp or llama.cpp?
The text was updated successfully, but these errors were encountered: