-
Notifications
You must be signed in to change notification settings - Fork 488
Increasingly occupied GPU memory #108
Comments
I guess it may relate to RNN compact weight, but I don't know how to fix it. |
Hi, Have you found a fix to this? I am having a similar issue. It was working on pytorch 0.4.1, the compact weight warning was displayed once at the beginning only, and it continues normally until the end of the training. Thanks a lot. |
Also having this issue. I don't think it's related to the flatten_parameters() warnings. It seems to be correlated with the optimizer--specifically, the memory usage only starts to increase after it is switched to ASGD. |
@rewicks good call, the memory usage increases only with the ASGD optimizer. I think I have found the problem with it, but I am not sure how to solve it. I printed the tensors living in memory using the GPU memory profiling code mentioned at https://discuss.pytorch.org/t/how-to-debug-causes-of-gpu-memory-leaks/6741/3 , and used the PyCharm debugger to see the variables during training. The ASGD
As the epochs go on and on,
Should we change the |
I have found a solution. If it works for others as well, this issue can be closed. I have modified the ASGD optimizer using @mourga's port of AWD-LSTM for PyTorch 1.2.0, from: https://github.com/mourga/awd-lstm-lm In particular, in
|
hi, @AndreaLK3. It works for me as well. However, I don‘t achieve the perplexities that this instruction declares.
I just achieve perplexities of 64.74/62.23(validation/testing) with the same command. |
Hi~, I meet some problem while running your code on GPU. During training, the program will unexpectly consume GPU memory continuously, like from 2000 Mb -> 3000 Mb -> ... -> and finally explode. I use Python 3.6, Pytorch 0.4 and GPU with 12 GB memory.
The text was updated successfully, but these errors were encountered: