You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We should avoid specifying the dtype arg for any PyTorch functions, and instead let PyTorch figure out the dtype automatically. Only that way, AMP would be used. A specified dtype would overwrite AMP. Or if we have good reasons to specify the dtype, maybe only do this if AMP is disabled.
We should actually follow the dtype which is returned by the PyTorch function, and overwrite the dtype of our Tensor.
That's probably already it?
We should later test this on some common modules.
The text was updated successfully, but these errors were encountered:
Automatic mixed precision (AMP) in PyTorch, support for the PT engine (#1334, #1322).
We should make sure this properly works with the RETURNN frontend (#1120).
Small things we should take care of:
dtype
arg for any PyTorch functions, and instead let PyTorch figure out the dtype automatically. Only that way, AMP would be used. A specifieddtype
would overwrite AMP. Or if we have good reasons to specify thedtype
, maybe only do this if AMP is disabled.dtype
which is returned by the PyTorch function, and overwrite thedtype
of ourTensor
.That's probably already it?
We should later test this on some common modules.
The text was updated successfully, but these errors were encountered: