-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Question]The locally deployed deepseek-v3 loses 5 points compared to the API #212
Comments
Which inference engine are you using in the local deployment? |
@GeeeekExplorer vllm-0.6.6.post1 |
能分享一份本地部署的部署文档嘛 我也想本地部署一下试试 |
做了些集成,只能分享一些参数配置: |
请问8张H20速度每秒多少token |
@GeeeekExplorer 我试了vllm-0.6.6.post1 / vllm-0.6.6,都是53上下,但是API能跑到57左右。请问API和开源的模型是同一个模型吗? |
用两台H100吧,比H20快很多 |
I deploy deepseek-v3 locally using 8xH20, test LiveBench-0831 with temperature=0, and without system prompt. The result shows a 5-point drop compared to the API. Are this released model and the API the same model?
The text was updated successfully, but these errors were encountered: