-
Notifications
You must be signed in to change notification settings - Fork 1.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Lamma2 training on dataset downloaded from Huggingface. #3529
Comments
Hi @sudhir2016! Can you add this to the Ludwig config in the Colab notebook?
|
Than you @arnavgarg1. I did what you asked. Now I am getting this error. RuntimeError: Caught exception during model preprocessing: template invalid for zero-shot prompt: Prompt template must contain either the 'sample' field or one of the columns from the dataset. This is how the dataset I loaded from Haggingface using load_dataset looks like. DatasetDict({ |
Thanks @sudhir2016. Are you able to share your colab notebook with me? It might help me work through the issues so I can let you know what needed to be changed to get it to work. Based on your response and the error, I think there's 2 different things that need to be addressed:
Hope this helps unblock you, but also more than happy to look over your colab notebook. |
Sounds like there may be more than one issue here, but the need to manually specify the local backend is not desirable, so reverting the PR that caused this (#3531) and yanking v0.8.1. We'll follow-up with 0.8.2 soon, but in the meantime, I suggest downgrading to v0.8. |
Thank you @arnavgarg1. Here is the link to my Colab https://colab.research.google.com/drive/1Gu5l9Z-KQLaYWhPJqiW-bMLMF5GHTJnD?usp=sharing |
Thank you @tgaddair. Looking forward to further guidance. |
Hey @sudhir2016, thanks for sharing your notebook! Can you try converting the HF dataset into a Pandas DataFrame like this:
|
@arnavgarg1 we should add an issue to allow users to provide HuggingFace |
Thank you @tgaddair. I was able to train. I then tried to upload to hf_hub. But I am getting this error. Exception: Model artifacts not found at /content/results/model/model_weights. It is possible that model at '/content/results' hasn't been trained yet, or something wentwrong during training where the model's weights were not saved. The model weights directory contains three files -/content/results/api_experiment_run/model/model_weights/README.md
Please suggest. |
I thought that there was some issue with my training dataset so I trained on "ludwig://alpaca" dataset and I got exactly the same error when I tried to upload tp hf_hub. Exception: Model artifacts not found at /content/results/model/model_weights. It is possible that model at '/content/results' hasn't been trained yet, or something wentwrong during training where the model's weights were not saved. |
Hi @sudhir2016, sorry you're running into issues with uploading your model weights out of colab and into HuggingFace Hub. Give me a few hours to test this flow myself and come back with a solution. Are you able to share the exact command you're running to upload these model weights to HuggingFace Hub? |
Hi @sudhir2016, I was able to take a look at this and it seems like things are working fine. Here are some instructions on how to upload the model weights successfully (cc: @tgaddair let's update your colab notebook to also include these steps):
Let me know if this works. |
Thank you @arnavgarg1 for your detailed feedback. I am traveling so will try and revert after a few days |
I am trying to use Ludwig to finetune Llama2 7b in Colab with a Alpaca type dataset downloaded from Huggingface. Using the example notebook provided. Tried various approaches without success. Getting error LLM with quantization requires the 'local' backend to be set in the config. Please advise.
The text was updated successfully, but these errors were encountered: