Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

DDP has slower training speed #124

Open
jyuntins opened this issue May 28, 2024 · 3 comments
Open

DDP has slower training speed #124

jyuntins opened this issue May 28, 2024 · 3 comments

Comments

@jyuntins
Copy link

Hi, I am trying to Finetune HMR2.0 on 3 RTX3090. It works fine when I only train it on 1 gpu.
When I set trainer.devices=3, I received the error:
ValueError: ctypes objects containing pointers cannot be pickled

I found a workaround to solve this is to use ddp strategy just as you did in the ddp trainer config file.
However, when I set the trainer to ddp, the speed drops from 2.5 it/s to 7s/it.

I wonder which config file is used when you trained hmr2.0? Is there anyway to accelerate the training speed in this configuration?

@wtx9527
Copy link

wtx9527 commented Aug 2, 2024

Hi, I have the same problem, have you solved it?

@jyuntins
Copy link
Author

jyuntins commented Aug 2, 2024

Hi, @wtx9527
No, I didn't solve it.

@wangsen1312
Copy link

I met the same problem when I use the multi GPU training

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants