-
Notifications
You must be signed in to change notification settings - Fork 279
Add an option to use fp8-all-gather only without fp8 computation. #1093
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1093
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5c5f4f2 with merge base ae77f40 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This pull request was exported from Phabricator. Differential Revision: D63056142 |
CI error looks relevant. maybe take a look before landing?
|
7857bb1
to
2950a2e
Compare
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
This pull request was exported from Phabricator. Differential Revision: D63056142 |
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
2950a2e
to
0f3ef23
Compare
This pull request was exported from Phabricator. Differential Revision: D63056142 |
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
0f3ef23
to
5c5f4f2
Compare
This pull request was exported from Phabricator. Differential Revision: D63056142 |
…training Summary: In #1093 we added a config option, off by default, to use only float8 all-gather for training and do the matrix multiply in high precision. This seems generally useful for communication bound workloads, but we can probably think of a cleaner way to add this functionality (such as a weight wrapper tensor subclass). The current implementation adds non-trivial complexity and doesn't jive well with where we want to take this codebase. Since no one is using this internally or externally yet and we haven't talked about it in the release notes, I think we should do a BC-breaking delete as a one-off. However, if people have concerns - let me know and we can talk about less aggressive options. Test Plan: ``` ./test/float8/test_everything.sh ``` Reviewers: Subscribers: Tasks: Tags:
…training Summary: In #1093 we added a config option, off by default, to use only float8 all-gather for training and do the matrix multiply in high precision. This seems generally useful for communication bound workloads, but we can probably think of a cleaner way to add this functionality (such as a weight wrapper tensor subclass). The current implementation adds non-trivial complexity and doesn't jive well with where we want to take this codebase. Since no one is using this internally or externally yet and we haven't talked about it in the release notes, I think we should do a BC-breaking delete as a one-off. However, if people have concerns - let me know and we can talk about less aggressive options. Test Plan: ``` ./test/float8/test_everything.sh ``` Reviewers: Subscribers: Tasks: Tags:
#1451) for now, delete the float8-all-gather-only functionality from float8 training Summary: In #1093 we added a config option, off by default, to use only float8 all-gather for training and do the matrix multiply in high precision. This seems generally useful for communication bound workloads, but we can probably think of a cleaner way to add this functionality (such as a weight wrapper tensor subclass). The current implementation adds non-trivial complexity and doesn't jive well with where we want to take this codebase. Since no one is using this internally or externally yet and we haven't talked about it in the release notes, I think we should do a BC-breaking delete as a one-off. However, if people have concerns - let me know and we can talk about less aggressive options. Test Plan: ``` ./test/float8/test_everything.sh ``` Reviewers: Subscribers: Tasks: Tags:
#1451) for now, delete the float8-all-gather-only functionality from float8 training Summary: In #1093 we added a config option, off by default, to use only float8 all-gather for training and do the matrix multiply in high precision. This seems generally useful for communication bound workloads, but we can probably think of a cleaner way to add this functionality (such as a weight wrapper tensor subclass). The current implementation adds non-trivial complexity and doesn't jive well with where we want to take this codebase. Since no one is using this internally or externally yet and we haven't talked about it in the release notes, I think we should do a BC-breaking delete as a one-off. However, if people have concerns - let me know and we can talk about less aggressive options. Test Plan: ``` ./test/float8/test_everything.sh ``` Reviewers: Subscribers: Tasks: Tags:
Summary:
The implementation reuses
WeightWithDynamicFloat8CastTensor
class and theFloat8Linear
module.I added an if-else branch in the existing
Float8Linear
module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather.Differential Revision: D63056142