I am an AI infrastructure software engineer, and my current focus is on pgvecto.rs, a vector database implemented in Postgres.
Nothing here yet.
Thanks for the question. You need to store the full-precision vector data , but you can build the index with binary vectors. The memory usage of the index can be reduced. Data and index, are two different things that need storage.
No, that is actually a new feature of the OpenAI embedding model. You have the ability to selectively discard or drop certain dimensions, and the model will still function appropriately. Under the hood is Matryoshka Representation Learning https://aniketrege.github.io/blog/2024/mrl/