-
Notifications
You must be signed in to change notification settings - Fork 18.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Provide an Adam solver #2827
Comments
I would like to try to do this, if no one else has started. I took the time to update the adam implementation for Lasagne and torch to the latest version of the paper. So next I will try to implement it in caffe. |
Just started to write the solver Not sure if we really want to use
for computing the effective stepsize. |
Ok I was thinking perhaps: int t = this->iter_ / this->param_.stepsize() + 1; |
Thanks for the suggested solver. Closing as addressed by #2856 -- see the PR for status. |
Are there any plans to add Adam solver to Caffe in the near future?
Adam: A Method for Stochastic Optimization
Python implementation (Theano)
Lua implementation (Torch)
I have been using the Lua implemenation for training neural networks for low-level image processing (such as noise reduction and super-resolution) with great success, but could not reproduce results of similar quality in Caffe with SGD/Adagrad/Nesterov. There are reports of better results (faster, more accurate) when training various types of deep networks using Adam.
The text was updated successfully, but these errors were encountered: