Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature] [OIM] Make inference backend configurable #1525

Open
Yu-amd opened this issue Feb 11, 2025 · 0 comments
Open

[Feature] [OIM] Make inference backend configurable #1525

Yu-amd opened this issue Feb 11, 2025 · 0 comments
Assignees
Labels
feature New feature or request

Comments

@Yu-amd
Copy link

Yu-amd commented Feb 11, 2025

Priority

P2-High

OS type

Ubuntu

Hardware type

GPU-AMD

Running nodes

Single Node

Description

Based on discussion with Zhiwei at Intel, I understand there are plans to create OIM for selecting inference backend based on the models.

I'm opening this feature request to add the requirement that OIM should allow user to configure and override the backend used. The reason is that not all inference backends will be available on all hardware and software platforms due to various reasons, so we want to ensure the user can manually configure the actual inference backend for the OPEA workload.

@Yu-amd Yu-amd added the feature New feature or request label Feb 11, 2025
@Yu-amd Yu-amd changed the title [Feature] Make inference backend configurable [Feature] [OIM] Make inference backend configurable Feb 11, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feature New feature or request
Projects
None yet
Development

No branches or pull requests

2 participants