Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Deploying this fine-tuned model #11

Open
babla9 opened this issue Jul 6, 2024 · 3 comments
Open

Deploying this fine-tuned model #11

babla9 opened this issue Jul 6, 2024 · 3 comments

Comments

@babla9
Copy link

babla9 commented Jul 6, 2024

Thanks so much for your work on this!

How can I deploy this fine-tuned model (expose via API endpoint)? Can I use vLLM or a library like this: https://github.com/EricLBuehler/mistral.rs, which supports Phi3-Vision?

Thanks!

@2U1
Copy link

2U1 commented Jul 15, 2024

@babla9 vLLM and mistral.rs both supports phi3-vision.
But, merging the lora weights should be more easier, when you use it.

@kevintee
Copy link

@2U1 What do you mean by merging the lora weights? Lora for phi-3 vision is currently not supported in vLLM.

@2U1
Copy link

2U1 commented Jul 20, 2024

@kevintee When you fine tune the model with the code. You will get a adapter weight(Because it finetunes phi3 with lora). You should merge in to the original model to use it with vLLM.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants