-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[DPO] Merge initial peft model if trainer has a peft_config #956
Conversation
Co-authored-by: Shoaib Burq <saburq@gmail.com>
The documentation is not available anymore as the PR was closed or merged. |
@Elfsong can you check if i am testing the failure case you have in the test? |
I’m currently working on this demo FYI: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM! Thank you!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM - given that this is a common scenario (people can SFT their model with peft and use that model for DPO) ! Thanks @kashif !
…ace#956) * failing test Co-authored-by: Shoaib Burq <saburq@gmail.com> * merge initial peft model
fixes #742
We merge the initial peft adaptor if we are training with a peft_config