Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch_size is 16 or 64? #15

Open
qilong-zhang opened this issue Jul 12, 2021 · 1 comment
Open

Batch_size is 16 or 64? #15

qilong-zhang opened this issue Jul 12, 2021 · 1 comment

Comments

@qilong-zhang
Copy link

qilong-zhang commented Jul 12, 2021

Hi @TACJu, I notice you apply DDP with 4 GPUs in train.py. Therefore, if the batch_size in args is set to 16, then the overall batch_size will be 16x4=64.
However, in your paper, you say that the batch_size is 16. I also try batch_size 16x4 on Tesla V100, but OOM will be raised, so I wonder batch_size is 16 means 16 or 64? thanks!

@slothfulxtx
Copy link

slothfulxtx commented Sep 21, 2021

Maybe the batch_size is 16*4 = 64. I run the code with batch_size=4*4 , and the accuracy on the CUB_200_2011 dataset is only 90.9%. After changing the batch_size to 4*8 (limited by memory, 4*16 cause OOM on my server with 4 RTX3090 GPUs), the accuracy raises up to 91.4%.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants