-
Notifications
You must be signed in to change notification settings - Fork 3.6k
Impr docs on batch sizes and limits in distributed #21070
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
Borda
merged 10 commits into
Lightning-AI:master
from
SkafteNicki:docs_clarify_effective_batch_size
Sep 2, 2025
Merged
Impr docs on batch sizes and limits in distributed #21070
Borda
merged 10 commits into
Lightning-AI:master
from
SkafteNicki:docs_clarify_effective_batch_size
Sep 2, 2025
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
justusschock
approved these changes
Aug 14, 2025
Borda
approved these changes
Aug 18, 2025
Borda
reviewed
Sep 2, 2025
Borda
reviewed
Sep 2, 2025
Borda
reviewed
Sep 2, 2025
lantiga
approved these changes
Sep 2, 2025
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks good!
Borda
added a commit
that referenced
this pull request
Sep 3, 2025
* improve gpu_faq * add per device text * header * Apply suggestions from code review --------- Co-authored-by: Jirka Borovec <[email protected]> (cherry picked from commit d3996ad)
lantiga
pushed a commit
that referenced
this pull request
Sep 5, 2025
* improve gpu_faq * add per device text * header * Apply suggestions from code review --------- Co-authored-by: Jirka Borovec <[email protected]> (cherry picked from commit d3996ad)
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does this PR do?
Fixes #21023
Fixes #21022
The two issues linked was raised back-to-back basically contains the same confusion:
What happens in distributed settings with the batch size?
One is related to the effective batch size in distributed settings the other to
limit_*_batches
arguments. In both cases I have tried to make it clear that in distributed settings the values are not global but instead per device.Connected to this when i tried to look for the documentation on hardware acceleration I could not initially find it. It was due to me not connecting the headline
Interactive cloud development
to hardware acceleration. I guess it was originally named that to highlight lightning.ai as a service, but that is not mention it the docs. So the headline feels very confusing. If this is too much, I can revert this change.Before submitting
PR review
Anyone in the community is welcome to review the PR.
Before you start reviewing, make sure you have read the review guidelines. In short, see the following bullet-list:
Reviewer checklist
📚 Documentation preview 📚: https://pytorch-lightning--21070.org.readthedocs.build/en/21070/