Skip to content

Conversation

ervteng
Copy link
Contributor

@ervteng ervteng commented Mar 6, 2021

Proposed change(s)

LSTMs would store the memory as a tensor, and keep a gradient. So the next time around, Torch would try to backprop through this gradient. For long episodes, this caused the graph to explode.

Note that this does not affect previous releases of ML-Agents (R13 and earlier)

Types of change(s)

  • Bug fix
  • New feature
  • Code refactor
  • Breaking change
  • Documentation update
  • Other (please describe)

Checklist

  • Added tests that prove my fix is effective or that my feature works
  • Updated the changelog (if applicable)
  • Updated the documentation (if applicable)
  • Updated the migration guide (if applicable)

Other comments

@ervteng ervteng merged commit 347852b into main Mar 8, 2021
@delete-merged-branch delete-merged-branch bot deleted the develop-fix-lstm-torch-backprop branch March 8, 2021 17:42
@github-actions github-actions bot locked as resolved and limited conversation to collaborators Mar 8, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants