-
Notifications
You must be signed in to change notification settings - Fork 517
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Allow setting per-layer learning rates" #1612
Allow setting per-layer learning rates" #1612
Conversation
shaydeci
commented
Nov 6, 2023
•
edited
Loading
edited
- Added support for passing initial_lr as a dictionary.
- Removed the usages of update_param_groups and initialize_param_groups. Affected recipes had equivalent initial_lr mapping added to them (tested).
- For the edge cases of having an instantiated optimiser, I assign names to the parameter_groups and extrac an initial_lr mapping from them so our schedulers can be used.
…zer_initializer' into feature/SG-1209_introduce_optimizer_initializer
…zer_initializer' into feature/SG-1209_introduce_optimizer_initializer
src/super_gradients/recipes/training_hyperparams/coco2017_yolo_nas_train_params.yaml
Outdated
Show resolved
Hide resolved
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I went pretty brutal on this PR, sorry :)
I mean, it's good. It written well, it introduce new features while keeping support of existing logic, and it's clean. Just a few tricky things here and there.
And cherry on top - let's rename the PR title to be "Github release notes"-friendly. Something like: "Allow setting per-layer learning rates" |
…zer_initializer' into feature/SG-1209_introduce_optimizer_initializer
One last remark from me - please add integration test for any model of your choice where we train a model for one short epoch with backbone lr=0, nech=0.1xLR, head=LR to ensure all works end-to-end |
…zer_initializer' into feature/SG-1209_introduce_optimizer_initializer
Adde unit test training with frozen part of the net. |