-
Notifications
You must be signed in to change notification settings - Fork 8
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
"conf", "logits", and "probs" are all NaN #1
Comments
Hi, thanks for your interests! It looks like a NaN error occurs during training, could you check if the model's outputs and loss go NaN during training? I have previously experienced similar issues when training reward models with PEFT, there seems to be some issue with PEFT version and weights' precision that can cause NaN during training, you could also try adding |
Many thanks for your reply! |
The current linearised Laplace framework only looks at uncertainty on a single prediction token, this may not translate easily to sentence-level uncertainty, and computing Jacobians could be quite costly. Currently, we're only working on extending the single token uncertainty to other settings, e.g. reward modelling. Also we have released a new more efficient and self-contained library here, that we plan to maintain in the future. |
Thanks a lot! |
Hi! Thanks for your work.
I try to reproduce 'run_gpt.py' on several datasets.
However, the 'eval_res.json' shows: {"index": 0, "true": 0, "pred": 0, "conf": NaN, "logits": [NaN, NaN], "probs": [NaN, NaN]} from the first time save the checkpointss. I try different learning rates but the results are the same.
Foundation model is llama2-7b-chat.
Do you know the reason? Thanks!
The text was updated successfully, but these errors were encountered: