-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Validation loss on pretraining?[Feature request] #20
Comments
Thanks for the question. However, I do not have validation dataset incorporated during training. Feel free to try it by your own! |
Thanks for your reply! By the way do you have validation data in the finetuning stage? |
And I have two extra questions which I am confused with:
|
|
Thanks for your reply! I'd like to confirm again that for pretraining stage, do you freeze the LLM weights? |
For pretraining, I never freeze the LLM weights. |
feature
Hi, I am trying to redo the pretrain step as you described in the readme doc. The training loss converges pretty fast. I find the logs in wandb and it turned out to be only containing the training loss. I wonder if you could add other metrics, like validation loss and perplexity.
Thanks a lot!
The text was updated successfully, but these errors were encountered: