You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When doing multi-rank training, it appears that the diversity across ranks is lower than it should be when compared with a single rank training with the same global batch size. We would expect performance to be the same but multi rank takes much more time -- it must mean that there are repeats across the ranks.
The text was updated successfully, but these errors were encountered:
When doing multi-rank training, it appears that the diversity across ranks is lower than it should be when compared with a single rank training with the same global batch size. We would expect performance to be the same but multi rank takes much more time -- it must mean that there are repeats across the ranks.
The text was updated successfully, but these errors were encountered: