Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[docs] Performance docs refactor p.2 #26791

Merged
merged 9 commits into from
Oct 24, 2023

Conversation

MKhalusova
Copy link
Contributor

@MKhalusova MKhalusova commented Oct 13, 2023

This PR continues performance docs refactor in the transformers docs. It focuses mainly on the "Efficient Training on Multiple GPUs" doc and contains the following changes:

  • Improves clarity and readability
  • Adds links to Accelerate where relevant
  • Removes a duplicated chunk of content
  • Resolves some formatting issues

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Oct 13, 2023

The documentation is not available anymore as the PR was closed or merged.

@MKhalusova MKhalusova marked this pull request as ready for review October 13, 2023 13:16
Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks very good on my end, thanks a lot @MKhalusova !!

Copy link
Member

@stevhliu stevhliu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for your work on this massive doc! Some thoughts:

  1. I feel like the Concepts section is a bit redundant when we go into much more detail about these techniques in their individual sections. Maybe we can add these terms to the Glossary?
  2. From the DP + PP section to the FlexFlow section, none of these techniques are supported or integrated in Transformers, so I'm not really sure we need to add these here. I feel like it would be frustrating for a user coming to the doc expecting to use DP + PP only to see that it's not implemented. I think putting them into a new "Parallelism" doc in the conceptual guides section may be better.

docs/source/en/perf_train_gpu_many.md Outdated Show resolved Hide resolved
docs/source/en/perf_train_gpu_many.md Outdated Show resolved Hide resolved
docs/source/en/perf_train_gpu_many.md Outdated Show resolved Hide resolved
docs/source/en/perf_train_gpu_many.md Outdated Show resolved Hide resolved

1. DP+PP+TP+ZeRO-1
🤗 Transformers status: not yet integrated. We already have our models FX-trace-able via [transformers.utils.fx](https://github.com/huggingface/transformers/blob/master/src/transformers/utils/fx.py), which is
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think it's cool to mention FlexFlow, but if it isn't integrated in Transformers then I'm not sure we need to add this section here because it doesn't really help users who want to try it out but it's not available.

Also not really a fan of "so someone needs to figure out what needs to be done to make FlexFlow work with our models". The tone sounds a bit demanding and not as friendly (totally understand this was already here before you started working on it 🙂 ).

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Rephrased the sentence. I'm on the fence about removing the content, or moving it to another section.

MKhalusova and others added 3 commits October 23, 2023 11:25
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Copy link
Collaborator

@ArthurZucker ArthurZucker left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Great work! 🤗

docs/source/en/glossary.md Show resolved Hide resolved
docs/source/en/perf_train_gpu_many.md Show resolved Hide resolved
docs/source/en/perf_train_gpu_many.md Outdated Show resolved Hide resolved
@MKhalusova MKhalusova merged commit 9333bf0 into huggingface:main Oct 24, 2023
8 checks passed
staghado pushed a commit to staghado/transformers that referenced this pull request Oct 24, 2023
* initial edits

* improvements for clarity and flow

* improvements for clarity and flow, removed the repetead section

* removed two docs that had no content

* Revert "removed two docs that had no content"

This reverts commit e98fa2f.

* Apply suggestions from code review

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>

* feedback addressed

* more feedback addressed

* feedback addressed

---------

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
i4never pushed a commit to i4never/transformers that referenced this pull request Oct 25, 2023
* initial edits

* improvements for clarity and flow

* improvements for clarity and flow, removed the repetead section

* removed two docs that had no content

* Revert "removed two docs that had no content"

This reverts commit e98fa2f.

* Apply suggestions from code review

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>

* feedback addressed

* more feedback addressed

* feedback addressed

---------

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
EduardoPach pushed a commit to EduardoPach/transformers that referenced this pull request Nov 19, 2023
* initial edits

* improvements for clarity and flow

* improvements for clarity and flow, removed the repetead section

* removed two docs that had no content

* Revert "removed two docs that had no content"

This reverts commit e98fa2f.

* Apply suggestions from code review

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>

* feedback addressed

* more feedback addressed

* feedback addressed

---------

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants