Full Jax/NNX implementation of microdiffusion (Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget). Adapted from existing official(Sony Research) and unofficial(SwayStar123) implementations.
Training/experimentations are proudly sponsored by Google TRC(TPU Research Cloud) grant.
microdit_trainer.py contains the complete(and wacky) single-file code for training a MicroDiT model from scratch,
on the common-canvas dataset. (you can also specify a smaller split of the ~15M images in the config
class).
Apache 2.0 License of course :)
@article{Sehwag2024MicroDiT,
title={Stretching Each Dollar: Diffusion Training from Scratch on a Micro-Budget},
author={Sehwag, Vikash and Kong, Xianghao and Li, Jingtao and Spranger, Michael and Lyu, Lingjuan},
journal={arXiv preprint arXiv:2407.15811},
year={2024}
}