-
Notifications
You must be signed in to change notification settings - Fork 3.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Deprecate the nvidia/apex integration #14416
Comments
I'm in favor |
deepspeed still supports amp, so we should also update the |
also, do you know if PyTorch is working on O1/O3 support natively? |
I wonder if DeepSpeed is impacted with the same checkpointing problems when apex is used.
I don't think so. It might not provide relevant efficiency improvements for most use cases so maybe they scrapped supporting it. @ptrblck, if you have any insights here, we'd love to hear them from you 🙇 Also relevant: pytorch/pytorch#52279 |
The native amp implementation via |
Proposed refactor
Deprecation:
ApexMixedPrecisionPlugin
and passingTrainer(amp_backend=...)
. To be removed in 1.10Removal:
Motivation
APEX AMP can be regarded as deprecated in favor of PyTorch AMP which Michael Carilli implemented and advocated in #1337.
Most developer activity in the nvidia/apex repository happen in either apex/transformer, apex/optimizers, tests/L0, and/or apex/contrib. apex/amp directory hasn't seen changes for about 2 years
Given the 2-year hibernation would make it almost impossible to resume the support for the different optimization levels to O2.
It's unclear whether any nvidia teams use our apex plugin internally.
And the nvidia team is unable to provide support for apex bugs.
If you enjoy Lightning, check out our other projects! ⚡
Metrics: Machine learning metrics for distributed, scalable PyTorch applications.
Lite: enables pure PyTorch users to scale their existing code on any kind of device while retaining full control over their own loops and optimization logic.
Flash: The fastest way to get a Lightning baseline! A collection of tasks for fast prototyping, baselining, fine-tuning, and solving problems with deep learning.
Bolts: Pretrained SOTA Deep Learning models, callbacks, and more for research and production with PyTorch Lightning and PyTorch.
Lightning Transformers: Flexible interface for high-performance research using SOTA Transformers leveraging PyTorch Lightning, Transformers, and Hydra.
cc @tchaton @rohitgr7 @carmocca @justusschock @awaelchli @akihironitta @kaushikb11 @Borda
The text was updated successfully, but these errors were encountered: