Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[FEATURE] Faiss online training to quantize fp32 vectors as int8 vectors #1723

Open
naveentatikonda opened this issue May 30, 2024 · 0 comments · May be fixed by #2425
Open

[FEATURE] Faiss online training to quantize fp32 vectors as int8 vectors #1723

naveentatikonda opened this issue May 30, 2024 · 0 comments · May be fixed by #2425
Assignees

Comments

@naveentatikonda
Copy link
Member

naveentatikonda commented May 30, 2024

Is your feature request related to a problem?
Like the Lucene Scalar Quantizer, add support for Faiss engine which accepts fp32 vectors as input without any range limitations and dynamically scalar quantize (recompute the min-max or quantiles and requantize data if required while merging segments or if the range of new data that will be ingested lies outside of existing min-max or quantile range) them into int8 vectors using non-uniform/training mechanisms without the need for any external quantization or extra training steps.

Also, for the disk based vector search use Faiss as the default engine (by replacing lucene engine) for x4 compression to keep it consistent with other compression levels.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
Status: Now(This Quarter)
Development

Successfully merging a pull request may close this issue.

1 participant