-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dynamic/variable batch size support #16914
Comments
Hello |
Related to #15573 |
@HsunGong could you pls share your use case when these various batch sizes could be used? probably if you have a heterogenous set of GPU cards? |
For sure, we have 2080ti, 3090, a10, a40 gpu cards, and we want to put dynamic batch size according to the gpu. |
if not too late: I started developing dynamic batch size and the corresponding LR scaling on deepspeed in PR 5237. Hopefully it will be enabled with a simple config file change and work out of the box in lightning. Stay tuned, should be done this week. |
Description & Motivation
Support like:
GPU1: batch-size=24
GPU2: batch-size=12
GPU3: batch-size=16
GPU4: batch-size=24
As different batch has different length
See
microsoft/DeepSpeed#1051
https://github.com/facebookresearch/fairseq/blob/b5a039c292facba9c73f59ff34621ec131d82341/fairseq/data/data_utils.py#L282
Pitch
No response
Alternatives
No response
Additional context
No response
cc @Borda
The text was updated successfully, but these errors were encountered: