-
Notifications
You must be signed in to change notification settings - Fork 507
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Integrate Lora fine-tuned model with HF #2025
Comments
Hi @GLorenzo679 thanks for creating this issue. We recommended the first path because it allows you to easily specify an HF model hub ID for the base model, and this is not very easy to save in the adapter config in the torchtune training loop (this is because |
Thanks for the fast reply.
I solved this by deleting the |
I have an update on this issue. |
related: #2048 |
@GLorenzo679, i agree 100%. I am working on fixing it, so this is invisible to the user. Sorry about that. Meanwhile, to unblock you, please do it manually and delete from the folder the other .safetensor files. You may have to delete the safetensors.index.json too. |
Thanks for pointing out that related issue! |
hey folks, PR is merged: #2074 Now it should be much easier to use vllm/huggingface. Instructions are in the or description We will update the docs soon. Let us know if you find any issues and thanks for your patience :). |
@felipemello1 Where in the README are the instructions? I don't see them. |
Sorry, I meant to say that they are in the pr description -.- |
Hello, I'm having some issues when using a Lora fine-tuned model with HuggingFace
from_pretrained()
.I saw some discussions in the issues, and this one #933 caught my attention.
It's suggested to do this in order to load the fine-tuned model:
Isn't it possible to load the model (
AutoModelForCausalLM.from_pretrained(model_id)
) from a previously downloaded model (the one that you get with tune download), instead of having to download it again from HF?If I try to load the previously downloaded model I get this error:
Looking at the code that generates the error, it seems like the transformers library is checking if peft the package is installed. If it is so, it looks for the adapter_config.json file (in my case the one generated with tune run) which is missing the
base_model_name_or_path
field.The text was updated successfully, but these errors were encountered: