Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multi-GPU training for DGL+ALIGNN #7372

Closed
knc6 opened this issue Apr 29, 2024 · 2 comments
Closed

Multi-GPU training for DGL+ALIGNN #7372

knc6 opened this issue Apr 29, 2024 · 2 comments

Comments

@knc6
Copy link

knc6 commented Apr 29, 2024

❓ Questions and Help

Hi,

Here is an issue regarding multi-GPU usage case for DGL for atomistic prediction in ALIGNN: usnistgov/alignn#90

We implemented the DDP feature, but still the RAM memory cost is a challenge. Any thoughts/suggestions how to tackle this issue efficiently?

@rudongyu
Copy link
Collaborator

rudongyu commented May 9, 2024

Hi @knc6 , original dataset in older dgl put all features and graph structures in memory, and I guess your dataset is pretty large, which makes the RAM memory cost huge. You may try the graphbolt in the newest version of dgl: https://docs.dgl.ai/stochastic_training/ondisk-dataset.html. It supports on-disk storage of features.

@knc6
Copy link
Author

knc6 commented May 9, 2024

Thanks for the recommendation. I found that LMDB worked well dealing with large dataset.

@knc6 knc6 closed this as completed May 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants