Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Loss is not going down #33

Open
TerryMelody opened this issue Mar 17, 2024 · 4 comments
Open

Loss is not going down #33

TerryMelody opened this issue Mar 17, 2024 · 4 comments

Comments

@TerryMelody
Copy link

image
Hello dear authors! I trained ddim in another dataset. In 1200 epoches,, the loss still seems not going down constantly while sometimes loss became large. I wonder whether it is normal?

@heng0214
Copy link

heng0214 commented Apr 2, 2024

I have the same problem with you. Have you already solved this problem?

@yoshi-taka-desu
Copy link

we have this problem, too.
Could somebody tell me how to solve or possible causes?

@ashah01
Copy link

ashah01 commented Jun 29, 2024

This is a fairly standard phenomenon in diffusion training since the model samples timesteps uniformly in the forward pass. If the model has already seen and optimized a timestep several times, it will yield a lower loss score if presented with the same timestep later on compared to if it's presented with some other timestep it hasn't seen much before. One quick experiment you can implement to confirm this: try freezing the timestep vector and then train your model normally by iterating through your dataset. The loss should consistently go down.

@TerryMelody
Copy link
Author

This is a fairly standard phenomenon in diffusion training since the model samples timesteps uniformly in the forward pass. If the model has already seen and optimized a timestep several times, it will yield a lower loss score if presented with the same timestep later on compared to if it's presented with some other timestep it hasn't seen much before. One quick experiment you can implement to confirm this: try freezing the timestep vector and then train your model normally by iterating through your dataset. The loss should consistently go down.

Got it! Thanks for your reply!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants