You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Support merging AdaLoRA adapters with base model when :
base model is loaded in int8 .
Base model loaded in 4bit using QLoRA method.
Motivation
This is helpful when the goal is to merge adapter weights for faster inference with 8bit model and 4bit model inference
Reduces memory usage when applying AdaLORA modules with LLMs like Falcon, LLAMA2 etc..
Your contribution
I am highly interested to add this feature to PEFT. #875 and #851 PR's will be my reference for writing this PR.
Please suggests what unit tests needs to be added.
Also, provide your invaluable suggestions and pitfalls that I should avoid when implementing for AdaLoRA.
The text was updated successfully, but these errors were encountered:
Please suggests what unit tests needs to be added.
I think you can take inspiration from the two PRs you quoted for determining what tests to add.
Also, provide your invaluable suggestions and pitfalls that I should avoid when implementing for AdaLoRA.
Hard to anticipate this. I would suggest to just go ahead and start with 8bit and see how far you get. Create a draft PR as soon as you got something going and we can give you feedback based on that.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Feature request
Hi @BenjaminBossan ,@younesbelkada, @pacman100
Support merging AdaLoRA adapters with base model when :
Motivation
Your contribution
I am highly interested to add this feature to PEFT.
#875 and #851 PR's will be my reference for writing this PR.
Please suggests what unit tests needs to be added.
Also, provide your invaluable suggestions and pitfalls that I should avoid when implementing for AdaLoRA.
The text was updated successfully, but these errors were encountered: