You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm working on an instance of a FAISS index distributed across EC2 instances in AWS.
Each EC2 instance has 8Gb of ram and I have ~250 million vectors to index, with 128 dimensions each (after PCA).
In theory, 8Gb could fit: 8E9bytes / (128 dim/vector * 4 bytes/dim) ~= 15.6 million vectors but I assume the index itself has additional memory requirements, and the search operations themselves consume additional memory.
Are there any guidelines for how many "shards" I should distribute my index into?
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
I'm working on an instance of a FAISS index distributed across EC2 instances in AWS.
Each EC2 instance has 8Gb of ram and I have ~250 million vectors to index, with 128 dimensions each (after PCA).
In theory, 8Gb could fit:
8E9bytes / (128 dim/vector * 4 bytes/dim) ~= 15.6 million vectors
but I assume the index itself has additional memory requirements, and the search operations themselves consume additional memory.Are there any guidelines for how many "shards" I should distribute my index into?
Cheers!
Beta Was this translation helpful? Give feedback.
All reactions