You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
nice work to make swa work in object detection!
i have one question about same epoch level comparison.
the result looks like faster rcnn r50 1x + 1x swa extra training get same result as faster rcnn r50 2x?
i think maybe some problem.
faster rcnn r50 1x + 1x swa extra training use cyclic training, but origin faster rcnn r50 2x use step down lr training.
this mismatch may lead to differenct converge. i think the best way is to train models from scratch with cyclic training to get a fair comaprison.
swa needs to change batch norm param to match average weight. frozen bn may harm the final ensemble result
The text was updated successfully, but these errors were encountered:
I think it is normal to have the same results. Because 1x training is actually not enough to achieve the performance saturation point in this case. What you should expect for swa training is if it can further improve the saturation performance.
You can try training Faster RCNN from scratch with cyclic learning rates and probably you get slight worse results. Do not ask me why I know this.
The reason for freezing BN in backbone is that the batch size used in training object detectors is not large enough to compute accurate statistics for BN. There are experiments in Section 5.2 of the MMDetection paper. In practice, there is still a considerable AP improvement with frozen BN. So it should not be a big problem.
nice work to make swa work in object detection!
i have one question about same epoch level comparison.
the result looks like faster rcnn r50 1x + 1x swa extra training get same result as faster rcnn r50 2x?
i think maybe some problem.
this mismatch may lead to differenct converge. i think the best way is to train models from scratch with cyclic training to get a fair comaprison.
The text was updated successfully, but these errors were encountered: