Replies: 1 comment 2 replies
-
Batch size shouldn't affect k-means memory consumption. It is only used by the Tester class for computing the embeddings.
Does it use the GPU even though you've specified CPU?
So it takes 80GB of memory using 64-dim embeddings? How big is your dataset? Could you see how much memory it takes with 16-dim embeddings? |
Beta Was this translation helpful? Give feedback.
2 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi Kevin, Thanks for creating this wonderful library! I have been using your library without problem before, but recently I seemed to be running into (GPU) OOMs (although I have a 32GB V100). Basically, the Jupyter notebook would always die at validation (k-means) time of the first epoch here:
I've already set both train and test batch sizes to 4, and output dimension of embedder to 16, which is obviously pretty small. I also tried to use CPU-only by reinstalling FAISS-cpu (naively setting
os.environ['CUDA_VISIBLE_DEVICES']='-1'
anddevice=torch.device('cpu')
does not work), and it worked smoothly with batch size as large as 64 and output dimension as large as 64 (80GB RAM). I would greatly appreciate any advice/suggestions!Beta Was this translation helpful? Give feedback.
All reactions