Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Remove device_map when training 4,8-bit model. #534

Merged

Conversation

SunMarc
Copy link
Member

@SunMarc SunMarc commented Jun 1, 2023

This PR fixes the examples using 4-bit and 8-bit model when training. It removes the device_map arg as it is not needed anymore after this PR. You should use device_map = 'auto' only for inference.

@review-notebook-app
Copy link

Check out this pull request on  ReviewNB

See visual diffs & provide feedback on Jupyter Notebooks.


Powered by ReviewNB

@HuggingFaceDocBuilderDev
Copy link

HuggingFaceDocBuilderDev commented Jun 1, 2023

The documentation is not available anymore as the PR was closed or merged.

Copy link
Contributor

@younesbelkada younesbelkada left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks @SunMarc for cleaning this up! 🚀

Copy link
Contributor

@pacman100 pacman100 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thank you @SunMarc for updating all the examples to reflect the important change of not having to specify device_map with quantization when training

@pacman100 pacman100 merged commit fcff23f into huggingface:main Jun 2, 2023
@SunMarc SunMarc deleted the update_quantized_model_training_examples branch June 2, 2023 13:24
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants