Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

batch_size and the unstable loss #5

Open
SkySeaWalk opened this issue Jun 6, 2020 · 1 comment
Open

batch_size and the unstable loss #5

SkySeaWalk opened this issue Jun 6, 2020 · 1 comment

Comments

@SkySeaWalk
Copy link

@lzx551402 @zjhthu Hi, thanks a lot for your great work.
I'v trained aslfeat with circle loss , and when I use tensorboard to see the loss , I found it decreased unstably from step 5K to 400K. I also check the default batch_size is 2.

  1. Did you do some experiments on the batch_size or other hyperparameters?
  2. Would the loss be more stable, if I increase the batch_size? or get a better model ?
  3. which base model did you use, when you train dcn with circle loss ? 380K step, 100K, or others?

Best Regards! look forward to your reply.

@EstherJinyan
Copy link

@lzx551402 @zjhthu Hi, thanks a lot for your great work.
I'v trained aslfeat with circle loss , and when I use tensorboard to see the loss , I found it decreased unstably from step 5K to 400K. I also check the default batch_size is 2.

  1. Did you do some experiments on the batch_size or other hyperparameters?
  2. Would the loss be more stable, if I increase the batch_size? or get a better model ?
  3. which base model did you use, when you train dcn with circle loss ? 380K step, 100K, or others?

Best Regards! look forward to your reply.

I have the same problem with you,have you solved this problem?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants