Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

question about Pretrained LLAMA applicable to Llama_adapter model. thanks #134

Open
jzssz opened this issue Nov 24, 2023 · 1 comment
Open

Comments

@jzssz
Copy link

jzssz commented Nov 24, 2023

  1. I noticed that the pretrained llama model in the code is in .pth format. So if I have a pretrained llama model in .bin format, can I use it as a Llama base model to train llama-adapter v2 (put it in "--llama_path") ?

  2. If I have a llama (7B) pre-trained with LORA (get a light-weight model like"adapter_model.bin"), can I use it as a Llama base model to train llama-adapter v2 (put it in "--llama_path") ?

thanks a lot

@jzssz jzssz changed the title question about pretrained llama(with LORA),thanks question about pretrained llama,thanks Nov 24, 2023
@jzssz jzssz changed the title question about pretrained llama,thanks question about Pretrained LLAMA applicable to Llama_adapter model. thanks Nov 24, 2023
@csuhan
Copy link
Collaborator

csuhan commented Nov 30, 2023

I guess your checkpoint is transformers-based, while our code only support the original llama format (https://github.com/facebookresearch/llama). So you need to download the original llama weights instead of transformers format.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants