You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on May 7, 2026. It is now read-only.
Is your feature request related to a problem? Please describe.
Memory usage can have a significant disk footprint when SotA embedding models are used, such as text-embedding-3-large (3072 dimensions) or Qwen3-Embedding-8B (4096 dimensions), while some other models reach 8192 dimensions.
Describe the solution you'd like
I want to benefit from LanceDB's implementation of the RaBitQ data quantization that could dramatically reduce disk footprint by sacrificing a small portion of recall accuracy.
Describe alternatives you've considered
Use embedding models with low dimensionality.
Is your feature request related to a problem? Please describe.
Memory usage can have a significant disk footprint when SotA embedding models are used, such as text-embedding-3-large (3072 dimensions) or Qwen3-Embedding-8B (4096 dimensions), while some other models reach 8192 dimensions.
Describe the solution you'd like
I want to benefit from LanceDB's implementation of the RaBitQ data quantization that could dramatically reduce disk footprint by sacrificing a small portion of recall accuracy.
Describe alternatives you've considered
Use embedding models with low dimensionality.
Additional context
https://www.lancedb.com/blog/feature-rabitq-quantization