-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[FSDP] Adding Native FSDP Strategy #12447
Merged
SeanNaren
merged 13 commits into
Lightning-AI:master
from
sisilmehta2000:strategy/fsdp_native
May 5, 2022
Merged
[FSDP] Adding Native FSDP Strategy #12447
SeanNaren
merged 13 commits into
Lightning-AI:master
from
sisilmehta2000:strategy/fsdp_native
May 5, 2022
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
sisilmehta2000
requested review from
tchaton,
SeanNaren,
carmocca,
Borda,
awaelchli,
justusschock,
kaushikb11,
williamFalcon and
rohitgr7
as code owners
March 24, 2022 21:19
akihironitta
added
feature
Is an improvement or enhancement
strategy: fairscale fsdp (removed)
Fully Sharded Data Parallel
labels
Mar 25, 2022
So some of the tests like conda(1.0, 1.8) are failing because it can't find torch.distributed.fsdp. That only exists in torch version 1.11. |
sisilmehta2000
force-pushed
the
strategy/fsdp_native
branch
from
March 25, 2022 01:52
18d8c65
to
0732774
Compare
kaushikb11
reviewed
Mar 25, 2022
rohitgr7
reviewed
Mar 25, 2022
four4fish
reviewed
Mar 25, 2022
sisilmehta2000
force-pushed
the
strategy/fsdp_native
branch
3 times, most recently
from
March 27, 2022 01:30
a064b70
to
7903625
Compare
carmocca
added
strategy: fsdp
Fully Sharded Data Parallel
and removed
strategy: fairscale fsdp (removed)
Fully Sharded Data Parallel
labels
Mar 28, 2022
zhaojuanmao
reviewed
Mar 28, 2022
sisilmehta2000
force-pushed
the
strategy/fsdp_native
branch
from
April 5, 2022 17:48
a1e038e
to
b77b121
Compare
sisilmehta2000
force-pushed
the
strategy/fsdp_native
branch
2 times, most recently
from
April 6, 2022 20:26
ef4c67e
to
118a0ea
Compare
SeanNaren
reviewed
May 4, 2022
SeanNaren
reviewed
May 4, 2022
SeanNaren
reviewed
May 4, 2022
SeanNaren
reviewed
May 4, 2022
mergify
bot
added
ready
PRs ready to be merged
and removed
has conflicts
ready
PRs ready to be merged
labels
May 4, 2022
ananthsub
approved these changes
May 4, 2022
Co-authored-by: ananthsub <[email protected]>
Borda
reviewed
May 5, 2022
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
goooo 🎉
SeanNaren
approved these changes
May 5, 2022
12 tasks
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
feature
Is an improvement or enhancement
ready
PRs ready to be merged
strategy: fsdp
Fully Sharded Data Parallel
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
What does this PR do?
Pytorch 1.11.0 makes torch.distributed.fsdp available for public use (PR #72084) https://github.com/pytorch/pytorch/releases/tag/v1.11.0
Currently there is a strategies/fully_sharded.py strategy that uses the fairscale implementation of fully sharded data parallel technique. With the introduction of fsdp in torch.distributed we want to create a new strategy called strategies/fully_sharded2.py
Reasons to add a new strategy:
Fixes #12334
Does your PR introduce any breaking changes? If yes, please list them.
None
Before submitting
PR review
Anyone in the community is welcome to review the PR.
Before you start reviewing make sure you have read Review guidelines. In short, see the following bullet-list:
Did you have fun?
Make sure you had fun coding 🙃