Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

slow training in single GPU #25

Open
guanamusic opened this issue Jul 14, 2021 · 1 comment
Open

slow training in single GPU #25

guanamusic opened this issue Jul 14, 2021 · 1 comment

Comments

@guanamusic
Copy link

Huge thanks for implement! I have a question regarding the training time in the single GPU you mentioned.
I did the same training procedure in batch size 96 on the RTX 2080Ti GPU as you did, but it took a lot longer than the training time you mentioned (12hrs to ~10k training iterations).
I have no idea the cause of this issue at all. Could you explain your training environment precisely?

Please refer to my working environment at the bottom.
Docker environment with
CUDA 10.1
cuDNN v7
ubuntu 18.04
python 3.8

@ali-z-khan
Copy link

@ivanvovk I'm facing the same problem. Have tried running it on an 8x A100 server too but it has taken me 9 days to reach 200 iterations. Any pointers would be greatly appreciated.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants