How to set learning rate for SimCLR with multiple GPU training #373
Answered
by
ananyahjha93
FrankXinqi
asked this question in
Q&A
-
Beta Was this translation helpful? Give feedback.
Answered by
ananyahjha93
Nov 18, 2020
Replies: 1 comment
-
@FrankXinqi For SimCLR with a batch size of 256, you can use the regular adam optimizer with learning rates 1e-4/1e-4. LARS is preferred for bigger batch sizes like 1024 and above. Also, try using the new updated simclr from the master branch, the online fine tuning is fixed. We'll soon have the imagenet weights in as well for SimCLR. Also, swav has the provision of a queue to run with a batch size of 256 and the authors have shown swav to be much more robust to small batch sizes. That is something you might wanna look at. |
Beta Was this translation helpful? Give feedback.
0 replies
Answer selected by
Borda
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
@FrankXinqi For SimCLR with a batch size of 256, you can use the regular adam optimizer with learning rates 1e-4/1e-4. LARS is preferred for bigger batch sizes like 1024 and above.
Also, try using the new updated simclr from the master branch, the online fine tuning is fixed.
We'll soon have the imagenet weights in as well for SimCLR.
Also, swav has the provision of a queue to run with a batch size of 256 and the authors have shown swav to be much more robust to small batch sizes. That is something you might wanna look at.