Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

AdEMAMix8bit settings for FLUX #2872

Open
pAInCREAT0R opened this issue Sep 30, 2024 · 0 comments
Open

AdEMAMix8bit settings for FLUX #2872

pAInCREAT0R opened this issue Sep 30, 2024 · 0 comments

Comments

@pAInCREAT0R
Copy link

I was excited to give this one a try based on the research paper. Testing a full checkpoint train in FLUX and was hoping to utilize this optimizers quicker convergence. However, I cannot seem to get it to operate within the constraints of 24GB VRAM (RTX4090). If you try to enable any of the block swap options, it errors as not available. Same for the paged version. Anyone have a config to utilize this on checkpoint training under 24gb vram? Am I missing an option? Is it supposed to work with block swap? I am not posting as an error because it could easily be my error. The people keeping this up are 10X smarter than I am on these things.

I have not tried it in LoRA training, but maybe it works better there? I welcome any suggestions. Training with ADAFACTOR now but convergence is slow even at a learning rate higher than I really want to be using.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant