-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The Implementation of AdaLoRA (ICLR 2023) #233
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is awesome 🔥. Well done @QingruZhang and Thank you for making AdaLoRA easy to use for the community 🤗. LGTM!
Left a few comments and suggestions. could you also run make style
and make quality
to fix the code quality CI.
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
Co-authored-by: Sourab Mangrulkar <13534540+pacman100@users.noreply.github.com>
The documentation is not available anymore as the PR was closed or merged. |
Hello @QingruZhang, I applied AdaLoRA to Whisper large fine-tuning, here is the wandb run
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you @QingruZhang for iterating, LGTM! 🤗
Hello @pacman100 , thanks for merging the commmits and running the test for AdaLoRA! Typically, we should set the initial budget as 1.5 times of final target budget and tune the budget schedule to have enough final fine-tuning steps to get the good performance. Please let me know if there are more experimental tests I need to do. Thanks agian for your help during this process! |
hi @pacman100 , do you also measure the peak GPU memory consumption, training time and other metrics? it would also be interesting to compare all these metrics, including the quality, when we set the budget as 1.5X or 2X. |
and how about run some test on the Llama 7B or 13B model? |
Dear PEFT maintainers,
This is Qingru Zhang, the author of "Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning" (ICLR 2023, please
see the link). We would like to submit this PR to intergrate AdaLoRA into PEFT. It was a great discussion with Sourab about the implementation of AdaLoRA and its intergration into PEFT. Thanks a lot for Sourab's comments and support during we prepare this PR. It would be great to have AdaLoRA available in PEFT! Please let me know in case of any questions about the impelmentation.
Thanks,
Qingru