Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix Documentation in FSDP and DeepSpeed Concept Guide #2725

Merged
merged 5 commits into from
May 1, 2024

Conversation

fabianlim
Copy link
Contributor

What does this PR do?

Address @stas00's comments that was omitted when we merged #2674

Before submitting

  • This PR fixes a typo or improves the docs (you can dismiss the other checks if that's the case).
  • Did you read the contributor guideline,
    Pull Request section?
  • Was this discussed/approved via a Github issue or the forum? Please add a link
    to it if that's the case.
  • Did you make sure to update the documentation with your changes? Here are the
    documentation guidelines, and
    here are tips on formatting docstrings.
  • Did you write any new necessary tests?

Who can review?

Anyone in the community is free to review the PR once the tests have passed. Feel free to tag
members/contributors who may be interested in your PR.

@stas00 @muellerzr

@fabianlim
Copy link
Contributor Author

@muellerzr the documentation build failed. I think its because now deepspeed is an optional dep now. It needs to be installed in the github workflow somehow.

@muellerzr
Copy link
Collaborator

To get around this instead of [utils.DummyOptim] just do [utils.deepspeed.DummyOptim]. There's no need for deepspeed to be installed

Copy link
Contributor

@stas00 stas00 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you for the follow up PR, @fabianlim!

I added a few small suggestions if they resonate.

docs/source/concept_guides/fsdp_and_deepspeed.md Outdated Show resolved Hide resolved
docs/source/concept_guides/fsdp_and_deepspeed.md Outdated Show resolved Hide resolved
@fabianlim
Copy link
Contributor Author

fabianlim commented May 1, 2024

@stas00 thank you for your suggestions I have integrated them.

To get around this instead of [utils.DummyOptim] just do [utils.deepspeed.DummyOptim]. There's no need for deepspeed to be installed

@muellerzr I tried this but it seemed I need to fix the find_object_in_package function in doc_builder. I have made a PR.

I have also added the links for the FSDP and DS docs 3b1948f. Hope they are added in appropriate places. If not please let me know.

Im still on the fence whether to put back the raise that I removed in 19cfab4. Decided to replace it after all.

@muellerzr
Copy link
Collaborator

@fabianlim can you rebase from main? We just merged something that should fix it!

@fabianlim
Copy link
Contributor Author

@muellerzr just rebased!

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

Copy link
Collaborator

@muellerzr muellerzr left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks a bunch for all your hard work!

@muellerzr muellerzr merged commit 2c76733 into huggingface:main May 1, 2024
23 checks passed
@fabianlim fabianlim deleted the fsdp_deepspeed2 branch May 2, 2024 01:05
luowyang pushed a commit to luowyang/accelerate that referenced this pull request May 2, 2024
* address part of stats comments

* automatically set sync_module_states if low_cpu_mem is set

* Apply suggestions from @stas00

Co-authored-by: Stas Bekman <[email protected]>

* add links from fsdp and deepspeed docs. fix deepspeed imports

* replace raise in accelerate.launch

---------

Co-authored-by: Stas Bekman <[email protected]>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants