-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
RuntimeError: CUDNN_STATUS_BAD_PARAM in loss.backward() #2
Comments
As the problem occurred in the backward step, the network should have finished forward step. So I think the problem might be out of memory. |
@cfzd
b1_grad_w and b1_grad_h should be exchanged in bottom1_grad initialization. |
@manmanCover Sorry, I have fixed the code. It's really confusing that I never met this problem and my experiment gains better performance. I think I need to release a benchmark soon. |
@cfzd Maybe your samples are all squares.
|
@manmanCover I tested my implementation with moderate and aggressive memory settings. I found slightly unbalanced memory consumption, but neither of my GPUs is fully occupied.
|
@cfzd Thank you for your test. I have checked that my input feature size is [32, 64, 128], how about yours? |
@cfzd By the way, is your implementation also use the sliding windows? It seems like not... |
@manmanCover |
I did't find any description of sliding window in the paper and I didn't see any reason of using sliding window, as it is an attention module. |
@cfzd yeah, adaptive pooling can also be a choice. The author of psanet said they use sliding windows with different input size (hszhao/PSANet#11 (comment)). |
Thank you for your wonderful code! Have you met this problem before and do you know how to solve it?
The text was updated successfully, but these errors were encountered: