You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be great to allow the Flux optimizers to have parameters that can be tweaked from the outside. As an example, I would like to have an SGD where I can tune the learning rate over time. I could do this by destroying the optimizer and recreating it every epoch, (which is fine for SGD) but that hurts performance with optimizers that keep running estimates of gradient magnitude such as ADAM or RMSProp.
The text was updated successfully, but these errors were encountered:
I am thinking of treating the optimizers as structs rather than functions. For example, ADAM will be a struct having learning rate, momentum, velocity etc. as its fields. And there can be an update function which takes this struct and performs the update, thereby updating the momentum and velocity. In this way, we'll be able to mutate the learning rate.
It would be great to allow the Flux optimizers to have parameters that can be tweaked from the outside. As an example, I would like to have an SGD where I can tune the learning rate over time. I could do this by destroying the optimizer and recreating it every epoch, (which is fine for SGD) but that hurts performance with optimizers that keep running estimates of gradient magnitude such as ADAM or RMSProp.
The text was updated successfully, but these errors were encountered: