You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Sep 6, 2022. It is now read-only.
So, I have tried to implement the model more or less "from scratch" on the basis of this repository.
For that I have implemented a training loop which I am executing eagerly to be able to debug.
However, in doing so I noticed that my loss is just jumping around weirdly and I still haven't figured out exactly why. In order to get insight I intended to take a closer look towards the rnnt_loss() function. While I executed some simple text-examples, I noticed that calling rnnt_losss()repeatedly on the same input, the loss is always different. But not just that: It's monotonically increasing.
So, I have tried to implement the model more or less "from scratch" on the basis of this repository.
For that I have implemented a training loop which I am executing eagerly to be able to debug.
However, in doing so I noticed that my loss is just jumping around weirdly and I still haven't figured out exactly why. In order to get insight I intended to take a closer look towards the
rnnt_loss()
function. While I executed some simple text-examples, I noticed that callingrnnt_losss()
repeatedly on the same input, the loss is always different. But not just that: It's monotonically increasing.The code I am running:
Output:
I am on
tensorflow==2.2.0
and I compiled the warp-transducer with GPU support.The text was updated successfully, but these errors were encountered: