-
Notifications
You must be signed in to change notification settings - Fork 545
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Learning Rate scheduler #121
Comments
Hi, This is so-called a LambdaLinearScheduler. If warm_up_steps is given by a value other than [0] (say [2500]), then this setting generates the constant learning rate of 1e-5 with warm-up period of epoch 0 to 2500 (linearly increasing from 1e-6 to 1e-5) and constant learning rate of 1e-5 all the way to the end. Please refer to the following code snippet: `### LambdaLinearScheduler.py Running command:CUDA_VISIBLE_DEVICES=0 torchrun --master_port 10000 LambdaLinearScheduler.pyimport torch class LambdaWarmUpCosineScheduler2:
class LambdaLinearScheduler(LambdaWarmUpCosineScheduler2):
def main(args):
if name == "main":
|
Hi,
Thanks for sharing this amazing work. I have been using this hugging face-based repo for fine-tuning - https://github.com/huggingface/instruction-tuned-sd but I have some queries regarding the learning rate scheduler.
In the paper, it was mentioned -
We use a learning rate of 10−4 (without any learning rate warm up).
Is it constant or some scheduler is used, I do see some config in
train.yaml
but can't understandIf it's being used, what would be the equivalent of this scheduler on diffusers.
The text was updated successfully, but these errors were encountered: