-
Notifications
You must be signed in to change notification settings - Fork 26.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[docs] Performance docs refactor p.2 #26791
[docs] Performance docs refactor p.2 #26791
Conversation
The documentation is not available anymore as the PR was closed or merged. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks very good on my end, thanks a lot @MKhalusova !!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks for your work on this massive doc! Some thoughts:
- I feel like the Concepts section is a bit redundant when we go into much more detail about these techniques in their individual sections. Maybe we can add these terms to the Glossary?
- From the DP + PP section to the FlexFlow section, none of these techniques are supported or integrated in Transformers, so I'm not really sure we need to add these here. I feel like it would be frustrating for a user coming to the doc expecting to use DP + PP only to see that it's not implemented. I think putting them into a new "Parallelism" doc in the conceptual guides section may be better.
|
||
1. DP+PP+TP+ZeRO-1 | ||
🤗 Transformers status: not yet integrated. We already have our models FX-trace-able via [transformers.utils.fx](https://github.com/huggingface/transformers/blob/master/src/transformers/utils/fx.py), which is |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think it's cool to mention FlexFlow, but if it isn't integrated in Transformers then I'm not sure we need to add this section here because it doesn't really help users who want to try it out but it's not available.
Also not really a fan of "so someone needs to figure out what needs to be done to make FlexFlow work with our models". The tone sounds a bit demanding and not as friendly (totally understand this was already here before you started working on it 🙂 ).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Rephrased the sentence. I'm on the fence about removing the content, or moving it to another section.
Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Great work! 🤗
* initial edits * improvements for clarity and flow * improvements for clarity and flow, removed the repetead section * removed two docs that had no content * Revert "removed two docs that had no content" This reverts commit e98fa2f. * Apply suggestions from code review Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * feedback addressed * more feedback addressed * feedback addressed --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* initial edits * improvements for clarity and flow * improvements for clarity and flow, removed the repetead section * removed two docs that had no content * Revert "removed two docs that had no content" This reverts commit e98fa2f. * Apply suggestions from code review Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * feedback addressed * more feedback addressed * feedback addressed --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
* initial edits * improvements for clarity and flow * improvements for clarity and flow, removed the repetead section * removed two docs that had no content * Revert "removed two docs that had no content" This reverts commit e98fa2f. * Apply suggestions from code review Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com> * feedback addressed * more feedback addressed * feedback addressed --------- Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
This PR continues performance docs refactor in the transformers docs. It focuses mainly on the "Efficient Training on Multiple GPUs" doc and contains the following changes: