-
Notifications
You must be signed in to change notification settings - Fork 3
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clarification on the provided training log file #11
Comments
In fact, we use gradient accumulation during the training process, so the equivalent batch size would be larger than the one inside the script. According to our experimental experience, the performance is stable when equivalent batch size is large than 500. You can try to adjust the step of gradient accumulation to get different equivalent batch sizes. |
Okay so you're saying I could increase the |
Unfortunately the performance was not as expected when training a model with the
I trained on a single GPU though, should I be adjusting the batch size or learning rate in any way to compensate? |
Sorry to bother you again @MCC-WH but I also have a question about Table 5 in your paper. What values of K and L did you use to compute the mAP for Affinity Feature (second row)? Also do you L2 normalize the affinity features before re-ranking? |
Hi @MCC-WH. First, thanks for making your training code publicly available. I'm trying to reproduce your training results and have some questions about the provided log file. The README states that it's from a 200 epoch schedule but it only seems to contain 100 epochs?
Also I'm wondering if the log is from the rSfM120k or AugrSfM120k experiment? With a batch size of 256 there should be 91642 / 256 = 358 batches for rSfM120k and 274926 / 256 = 1074 for AugrSfM120k. However the log file indicates that training was run on 765 batches. Perhaps another batch size was used?
The text was updated successfully, but these errors were encountered: