-
Notifications
You must be signed in to change notification settings - Fork 4.1k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training recipe?? #14
Comments
Hi, thanks for your interest! We will release the training code once the Hugging Face interface to LLaMA becomes stable (merged into main). Our fine-tuning procedure is standard and was performed with huggingface's trainer. You can see our hyperparameters here. |
Hi, do you fine-tune the model using Next Token Prediction as the pre-training? And how much CUDA memory does the training cost? Is it possible to do the training using only one 80G A100 GPU? |
Hi all, We have released the training code, see https://github.com/tatsu-lab/stanford_alpaca#fine-tuning. Please open a new issue for any further questions/concerns. |
@rtaori as far as I can tell from your code it looks like standard teacher forcing (aka next token prediction), is this accurate? |
The blog says training recipe is too released in the code, but I cannot find it. Can you update the repo with code used for training the model, along with required dependencies/guide, etc, to help us do the same, maybe with bigger models.
Thanks for this awesome repo.
The text was updated successfully, but these errors were encountered: