-
Notifications
You must be signed in to change notification settings - Fork 66
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
sh_albert_cls.sh get a weired score 0.499 #22
Comments
Which model do you use? Because I use RTX 3060 graphics card for training, I choose albert-base-v2, and the result I get is |
I also use albert-base-v2. After downsampling, the ratio of positive and negative samples is approximately equal to 40000:50000. In the first 3000 steps, the score is approximately equal to 0.71, the highest is 0.76, and then it starts to decline. What’s even stranger is that after I remove the downsampling, the score is no longer 0.499, but 0.61 |
我今天又把epoch改为5跑了下,结果也是 |
应该不需要吧,我执行 |
请问是如何进行下采样的 |
就是对负样本随机下采样 |
Hi i have same problem when run sh_albert_cls.sh, the score remains unchanged whatever the param is |
Have you solved it now? |
Unfortunately not yet. Hope author or anybody can help |
Dear, sorry for the late reply. It seems that albert-base-v2 does not converge according to your results. I did not try the base models, and the given hyper-parameters are tuned for xxlarge models. Maybe you can try to use the settings in google-research/albert#235 |
Has anyone ran sh_albert_cls.sh and got a strange score, keeping 0.499 unchanged
The text was updated successfully, but these errors were encountered: