Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

With the same experiment setting, the miss rate is not as good as that in the paper. #8

Open
sxlpris opened this issue Nov 22, 2018 · 2 comments

Comments

@sxlpris
Copy link

sxlpris commented Nov 22, 2018

I have trained the proposed model on 2 GTX 1080Ti GPUs (a mini-batch contains 10 images per GPU), which is the setting in the paper. And, I trained for 240k iterations totally, with the initial learning rate of 0.0001 in 160k iteration and degrade the initial learning rate to 0.00001 in the last 80k iterations, but the miss rate is 14.52%. Where might the problem arise?

@gittigxuy
Copy link

gittigxuy commented Jan 27, 2019

@sxl1995 ,could you please share your model with me?my email is:1262485779@qq.com

furthermore,how could you train ?do you use wma method?

@Jokoe66
Copy link

Jokoe66 commented Apr 29, 2019

I got the same problem. For the limit of GPU I set batch_size as 4 and train the model using 1 GTX 1080Ti. I trained the model for 150 epochs with init_lr of 1e-4, and chose the best performing one which is saved at about epoch 101. Then I degreed the init_lr by 10, i.e. 1e-5 and trained until converging. Finally I got best MR of around 14.59% on validation set. I wonder if there is any detail about the training process which I missed or if random inital weights do make some difference. @liuwei16 @VideoObjectSearch

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants