You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Could you please specify the environment in which these benchmark tests were conducted? (e.g., On-premises, OCI, AWS, Azure, etc.)
Regarding the H100 80G BF16 benchmark tests, what criteria were used to determine the input sequence length, MicroBatchSize, GradAccum, and GlobalBatchSize parameters?
For the same H100 80G BF16 tests, could you explain the rationale behind setting the Activation Checkpointing option to True?
I would greatly appreciate any insights you could provide on these matters. Thank you in advance for your time and assistance.
Additional context
The text was updated successfully, but these errors were encountered:
❓ Question
I am writing to request some clarification regarding the benchmark tests shared on GitHub at https://github.com/mosaicml/llm-foundry/blob/main/scripts/train/benchmarking/README.md.
Specifically, I have the following questions:
I would greatly appreciate any insights you could provide on these matters. Thank you in advance for your time and assistance.
Additional context
The text was updated successfully, but these errors were encountered: