You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Jul 24, 2024. It is now read-only.
I'm wondering if you have encountered this recurring spike in the training loss and why might this occur.
It does appear to occur upon processing the last batch but I don't think it has to do with uneven batch size because the last batch is dropped. The data is shuffled so it can't be any particular examples causing this. But it's not clear why the processing of the last batch would produce these spikes.
Thanks.
The text was updated successfully, but these errors were encountered:
@slala2121 I'm actually experiencing the exact same behavior, and I'm also shuffling + dropping the last batch. Did you happen to figure out the reason or resolve this?
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
I'm wondering if you have encountered this recurring spike in the training loss and why might this occur.
It does appear to occur upon processing the last batch but I don't think it has to do with uneven batch size because the last batch is dropped. The data is shuffled so it can't be any particular examples causing this. But it's not clear why the processing of the last batch would produce these spikes.
Thanks.
The text was updated successfully, but these errors were encountered: