Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MultiRC task- v1.0 results are better than v1.3 (Unknown reason, after reviewing changes) #1081

Open
jeswan opened this issue Sep 17, 2020 · 1 comment

Comments

@jeswan
Copy link
Contributor

jeswan commented Sep 17, 2020

Issue by varunchaudharycs
Wednesday Apr 29, 2020 at 06:22 GMT
Originally opened as nyu-mll/jiant#1081


Hi,
I'm trying out the MultiRC dataset task(Khasabi et. al 2018). Using the two jiant versions gave the following F1 scores(using BERT-base):
jiant v1.3 = ~58
jiant v1.0 = ~65
with both having the same configurations. I looked over the changes and found one:
v1.0 gives (para + ques + ans option) as BERT input
v1.3 gives ([CLS] para [SEP] ques + ans option [SEP]) as BERT input

I tried doing the same thing as v1.0 in v1.3 but observed no changes in the F1 score(rather F1 decreased).

Kindly let me know if I may try out some other changes between the two versions that I have missed earlier. This drastic drop in performance over version upgrade is hard to understand.

@jeswan
Copy link
Contributor Author

jeswan commented Sep 17, 2020

Comment by W4ngatang
Tuesday May 05, 2020 at 02:45 GMT


Hey Varun,

Sorry to hear that. So we have more info, can you share your config file? Also, has the drop been consistent across random seeds before and after the version change?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant