Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

EfficientVit Segmentation Hyper-Parameters #171

Open
davidhuangal opened this issue Jan 24, 2025 · 0 comments
Open

EfficientVit Segmentation Hyper-Parameters #171

davidhuangal opened this issue Jan 24, 2025 · 0 comments

Comments

@davidhuangal
Copy link

davidhuangal commented Jan 24, 2025

I would like to kindly ask that the training parameters used for training of the semantic segmentation EfficientViT models be released.

The paper only specifies that the AdamW optimizer was used and that a cosine learning rate decay was used, and I am having some trouble finding the right parameters to use to train the model well.

For example, it would be nice to know:

  • Batch size
  • Number of training iterations
  • Initial learning rate
  • If the learning rate was set differently for the backbone vs the SegHead
  • Settings for cosine lr scheduler. E.g., was a warmup used?
  • If weight decay was used
  • Augmentations used
  • Etc.

It would also be good to know if any of the hyper-params differed between training on ADE20K and training on Cityscapes.

Thank you for sharing this amazing work, it is so cool that such an elegant method can produce such powerful results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant