-
Notifications
You must be signed in to change notification settings - Fork 28.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
BART FP16 #3117
Comments
Not on my roadmap just yet, but I would definitely consider it if there were lots of demand. Since we only have inference code right now, the benefit seems marginal. |
@BramVanroy Should this issue be closed ? FP16 is not implemented yet. And the Keeping the issue open may make it easier for people to find it and show their potential interest in FP16. |
This should not be closed indeed. @sshleifer, we intend all the models to be compatible with FP16, this is the direction the field is going and with the Volta-level GPU being widespread now, there is less and less reason not to use mixed-precision fine-tuning (half memory and significantly faster). |
This can probably be fixed by changing the Do you mind fixing this and testing the failing script posted in the issue @sshleifer? |
Yep, on it! |
Hi, @sshleifer. Thank you so much for your effort on BART. I encountered the same fp16 issues today. The current BART code can be trained (without fp16) using the run_glue script in: https://github.com/huggingface/transformers/blob/master/examples/run_glue.py |
My bad, I thought @sshleifer's labeling was a note that he isn't planning to change anything |
No bad @sshleifer for the moment, please ping me with DM before adding "wontfix" labels to issues, thanks. |
🚀 Feature request
I would like to use BART in FP16 mode, but it seems impossible for now :
@sshleifer Do you plan to implement a FP16-friendly version of BART ?
The text was updated successfully, but these errors were encountered: