Skip to content
This repository has been archived by the owner on Jul 24, 2024. It is now read-only.

Spikes in training loss #189

Open
slala2121 opened this issue Feb 6, 2022 · 1 comment
Open

Spikes in training loss #189

slala2121 opened this issue Feb 6, 2022 · 1 comment

Comments

@slala2121
Copy link

I'm wondering if you have encountered this recurring spike in the training loss and why might this occur.

It does appear to occur upon processing the last batch but I don't think it has to do with uneven batch size because the last batch is dropped. The data is shuffled so it can't be any particular examples causing this. But it's not clear why the processing of the last batch would produce these spikes.

Thanks.

Screen Shot 2022-02-06 at 12 37 55 AM

@thecooltechguy
Copy link

@slala2121 I'm actually experiencing the exact same behavior, and I'm also shuffling + dropping the last batch. Did you happen to figure out the reason or resolve this?

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants