-
Notifications
You must be signed in to change notification settings - Fork 180
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add an option to use fp8-all-gather only without fp8 computation. #1093
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/1093
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 5c5f4f2 with merge base ae77f40 (): This comment was automatically generated by Dr. CI and updates every 15 minutes. |
This pull request was exported from Phabricator. Differential Revision: D63056142 |
CI error looks relevant. maybe take a look before landing?
|
7857bb1
to
2950a2e
Compare
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
This pull request was exported from Phabricator. Differential Revision: D63056142 |
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
2950a2e
to
0f3ef23
Compare
This pull request was exported from Phabricator. Differential Revision: D63056142 |
…torch#1093) Summary: The implementation reuses `WeightWithDynamicFloat8CastTensor` class and the `Float8Linear` module. I added an if-else branch in the existing `Float8Linear` module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather. Reviewed By: weifengpy Differential Revision: D63056142
0f3ef23
to
5c5f4f2
Compare
This pull request was exported from Phabricator. Differential Revision: D63056142 |
Summary:
The implementation reuses
WeightWithDynamicFloat8CastTensor
class and theFloat8Linear
module.I added an if-else branch in the existing
Float8Linear
module to re-use our existing logics to handle different casting cases, such as pre-/post-forward for delayed scaling, pre-compute amax for fp8-all-gather.Differential Revision: D63056142