-
Notifications
You must be signed in to change notification settings - Fork 204
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix documentation to explain how schedules work in optimizers. #399
Comments
copybara-service bot
pushed a commit
that referenced
this issue
Feb 5, 2024
The doc was slightly misleading see #399. PiperOrigin-RevId: 604372850
copybara-service bot
pushed a commit
that referenced
this issue
Feb 5, 2024
The doc was slightly misleading see #399. PiperOrigin-RevId: 604372850
copybara-service bot
pushed a commit
that referenced
this issue
Feb 6, 2024
The doc was slightly misleading see #399. PiperOrigin-RevId: 604372850
copybara-service bot
pushed a commit
that referenced
this issue
Feb 6, 2024
The doc was slightly misleading see #399. PiperOrigin-RevId: 604548137
Done in #778 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Discussed in #390
Originally posted by nalzok August 13, 2022
For example, in the documentation for
optax.adam
, we haveI noticed that
learning_rate
can be a function. How can a function be "a fixed global scaling factor"? What parameters should such a function take, and how does its return value affect the optimization process given that Adam has a learning rate schedule on its own?The text was updated successfully, but these errors were encountered: