Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Ensure that all operators properly support kAddTo in backward pass #9497

Open
asmushetzel opened this issue Jan 19, 2018 · 1 comment
Open

Comments

@asmushetzel
Copy link
Contributor

PR9495 fixes one operation where the backward pass of an operator did not obey the "req" argument properly (resp. not at all) and so ignored the "kAddTo" directive (which directs the operator to add the gradients to the output tensor(s) and not just assign them. This leads to wrong gradients in case that an operator fans out its output to more than one other operator.

We should examine all operators whether they all properly handle the "req" parameter in backward pass. There is at least one more that doesn't which is svm_output.

As this are basic and easy to fix but hard to detect problems (they may just slightly derail the training over time) we should really prioritize this sanity checking,

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Projects
None yet
Development

No branches or pull requests

2 participants