-
Notifications
You must be signed in to change notification settings - Fork 519
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot reproduce the result for bert-base-uncased
, avg_first_last
setting
#285
Comments
Hi, It looks like the dependency is the same as our experiment setting and the hardware shouldn't cause that much of a difference. Unfortunately I am also not sure what caused the difference.... have you tried testing the RoBERTa first-last avg? |
@gaotianyu1350 It turned out that due to the logic change in first-last avg pooling in this commit, it seems that the current codebase cannot reproduce the result for the models which use first-last avg (like BERT or RoBERTa). It would be very nice if you could add some notes in README or your paper about this discrepancy for those who are/will be trying to reproduce the result! :) |
Hi, Thanks for figuring it out!! Yeah it makes sense that using the contextual embedding improves the result. I'll add a note to the readme. |
Thank you for updating README! Closing this issue. |
@gaotianyu1350
Hi, thank you for the great work / publishing beautiful codes!
I have some questions reproducing the STS results for pre-trained bert models.
When I run the following command in my environment, I got higher STS scores comparing to the results shown in your paper.
Do you have any idea what is causing the issue?
Code executed
Results
Expected results (scores shown in your paper)
Strangely, I can fully reproduce the scores for SimCSE models via following command:
Here is the result of
pip freeze
and I am using one NVIDIA RTX 6000 Ada GPU.Thank you very much for your help!
pip freeze result
The text was updated successfully, but these errors were encountered: