You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Oct 16, 2023. It is now read-only.
Hi, @binmakeswell
Using enerogonAI insterad of colossal-AI should speed up inference on local machine with such as non-blocking pipeline parallel, redundant padding elimination, gpu offload, right?
If I does want to infer opt on local machine instead of http service, how should we modify the opt_server.py? Can you give us some examples?
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
hello,
I want to just inference of pre-trained model in the terminal, but I don't want to run a HTTP server. How could I do that?
The text was updated successfully, but these errors were encountered: