Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

difference between get_peft_model and from_pretrained #470

Closed
nuoma opened this issue May 19, 2023 · 3 comments
Closed

difference between get_peft_model and from_pretrained #470

nuoma opened this issue May 19, 2023 · 3 comments

Comments

@nuoma
Copy link

nuoma commented May 19, 2023

Hi, when I'm training LLAMA with Lora, i seem to get different results when loading Lora weights via get_peft_model and from_pretrained. Can't really tell why,both can infer successfully, but with vastly different results.

Could someone be so kind to tell me which is the correct way of doing this. Many thanks!

@yyqi17
Copy link

yyqi17 commented May 23, 2023

Hi nuoma, could you please show me your code in saving & loading checkpoints? I'm finetuning llama with lora, and seem to encounter some issues when infering my checkpoint (saved by model.save_pretrained() and loaded by PeftModel.from_pretrained(base_mode, lora_ckpt_path)). Thanks a lot!

@ingo-m
Copy link

ingo-m commented May 26, 2023

Possibly related? #503 🤔

@github-actions
Copy link

This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants