Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[low-bit optim] Fix Adam4bit support on PyTorch 2.3 and 2.4. Update AdamFp8 torch requirement #755

Merged
merged 7 commits into from
Sep 2, 2024

Conversation

gau-nernst
Copy link
Collaborator

Copy link

pytorch-bot bot commented Aug 27, 2024

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/ao/755

Note: Links to docs will display an error until the docs builds have been completed.

❌ 2 New Failures

As of commit d83a1c1 with merge base ba2d3b1 (image):

NEW FAILURES - The following jobs have failed:

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Aug 27, 2024
msaroufim
msaroufim previously approved these changes Aug 27, 2024
@gau-nernst gau-nernst marked this pull request as draft August 27, 2024 14:17
@gau-nernst
Copy link
Collaborator Author

convert this to draft since I'm also investigating torch version support of FP8 optim. FP8 optim has never run in CI due to sm89 constraint.

@gau-nernst
Copy link
Collaborator Author

gau-nernst commented Aug 27, 2024

Fixed issue with 4-bit Adam. Now 4-bit Adam works with PyTorch 2.3 likes in the past. Hopefully CI is green. The issue seems to be related to this pytorch/pytorch#128649

I kinda feel conflicted about this change, since now the optimizer state is flattened, instead of having the same shape as param. Will try a better solution in the future. I think it has to do with dynamic compile also. 4-bit optim is giving us a lot of headaches 🤣.

@msaroufim msaroufim self-requested a review August 27, 2024 15:12
@gau-nernst gau-nernst marked this pull request as ready for review August 27, 2024 15:13
@gau-nernst gau-nernst changed the title [4-bit optim] Update doc on torch version requirement [low-bit optim] Fix Adam4bit support on PyTorch 2.3 and 2.4. Update AdamFp8 torch requirement Aug 27, 2024
@msaroufim msaroufim merged commit 65f660d into pytorch:main Sep 2, 2024
15 of 17 checks passed
@gau-nernst gau-nernst deleted the adam4bit_doc branch September 2, 2024 19:25
jerryzh168 pushed a commit to jerryzh168/ao that referenced this pull request Sep 4, 2024
…damFp8 torch requirement (pytorch#755)

* update doc on torch version

* update doc

* update

* fix 4-bit problem

* update doc

* update
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed.
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants