Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about 'lr_mult' #11

Open
letitbehj opened this issue Apr 19, 2016 · 3 comments
Open

Question about 'lr_mult' #11

letitbehj opened this issue Apr 19, 2016 · 3 comments

Comments

@letitbehj
Copy link

letitbehj commented Apr 19, 2016

After reading the sample in http://caffe.berkeleyvision.org/tutorial/layers.html, I know 'lr_mult' is learning rate multipliers for the weights or the biases. But there are 3 'lr_mult' in deep_lstm_short.prototxt, what does the third lr_mult mean?
I am still a novice in caffe, sorry to disturb you.

@junhyukoh
Copy link
Owner

There are three weights in the lstm layer.
The first one corresponds to input-to-hidden weight
The second one corresponds to hidden-to-hidden weight
The third one corresponds to bias.
So, the third lr_mult is the lr multiplier for the bias.

@letitbehj
Copy link
Author

Thank U~

@erinchen824
Copy link

I found that there are 3 lstm layers with the first one have the lr_mult. Does it mean the latter two lstm layers won't get param updated? BTW, the lr_mult for conv layer is 1 and 2 for weight and bias respectively. Here, the lr_mult is 1,1,2 for the upper mentioned three condition. Is 1,1,2 make any sense? Thanks! @junhyukoh

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants