Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CUDAEXecutionProvider fails for BAAI/bge-m3 #16

Open
R4ZZ3 opened this issue Aug 22, 2024 · 1 comment
Open

CUDAEXecutionProvider fails for BAAI/bge-m3 #16

R4ZZ3 opened this issue Aug 22, 2024 · 1 comment

Comments

@R4ZZ3
Copy link

R4ZZ3 commented Aug 22, 2024

Hi,
I have been trying to get this to run on Runpod for few hours now with different image/template settings.
I have succesfully been able to run this using the torch backend but it seemed to run on CPU based on the runpod monitoring although GPU allocation showed that something was allocated to GPU.
I have tried both cuda 11.8 and 12.1 images. Both dev and stable. Here are my latest template configs:
image

image

@R4ZZ3
Copy link
Author

R4ZZ3 commented Aug 27, 2024

Nevermind, running locally with smaller max length + CudaOnnx solved the issue or now the performance is enough to run locally

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant