Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
feat(train): add optional
accumulate_grad_batches
config param (#306)
* feat(train): add accumulate_grad_batches hparam Add an `accumulate_grad_batches` param to the `train` part of the config to allow for gradient accumulation. This updates the gradients once every `accumulate_grad_batches` batches, with a default value of 1 to not break any existing configs. * fix(train): normalize loss when using gradient accumulation
- Loading branch information