Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Again: Training imagenet: loss does not decrease #3243

Closed
dereyly opened this issue Oct 23, 2015 · 4 comments
Closed

Again: Training imagenet: loss does not decrease #3243

dereyly opened this issue Oct 23, 2015 · 4 comments
Labels

Comments

@dereyly
Copy link

dereyly commented Oct 23, 2015

I have the same problem as #401
But have different story. Half year ago I sucsefully learned Imagenet caffe net with cuda 6.5 and cudnn_v2 (45000 iter with same resault as base model). And now i try to learn it again and failed -- loas doesn't decrease it is about 6.9 and accuracy =0. The sistem and driver is the same.
Test of old model is ok (expected accuracy).

I read the issues and there is no problem with shuffle the lmbd data is the same and same protobuff and build

Before write my question I rebuild new caffe with cuda 7.0 and cudnn_v3 and new driver 346.39 and have the same problem.

I run test about 10k-20k iterations. My old save have 16% acuracy on 10k iteratioons now its 0.

@baiyancheng20
Copy link

hi, have you resolved the problem? I met this problem, too.

@dereyly
Copy link
Author

dereyly commented Jan 19, 2016

Hi. Yes. I use MSRA init in convolution layers and same gaussian in full connected
But later I try LSUV: https://github.com/ducha-aiki/LSUVinit (All you need is good init)
it is better then MSRA
Finally batch normalization may solve this problem too

@baiyancheng20
Copy link

@dereyly thanks you. I use the MSRA init, and it works. Have you re-trained the alex network? Did it achieve the same test accuracy (58% on val)?

@dereyly
Copy link
Author

dereyly commented Jan 20, 2016

Yes. I train 450k (but u can stop earlier) iterations and get same result top1=58% and top5=81%

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

3 participants