You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi,
I'm trying out the MultiRC dataset task(Khasabi et. al 2018). Using the two jiant versions gave the following F1 scores(using BERT-base): jiant v1.3 = ~58 jiant v1.0 = ~65
with both having the same configurations. I looked over the changes and found one:
v1.0 gives (para + ques + ans option) as BERT input
v1.3 gives ([CLS] para [SEP] ques + ans option [SEP]) as BERT input
I tried doing the same thing as v1.0 in v1.3 but observed no changes in the F1 score(rather F1 decreased).
Kindly let me know if I may try out some other changes between the two versions that I have missed earlier. This drastic drop in performance over version upgrade is hard to understand.
The text was updated successfully, but these errors were encountered:
Comment by W4ngatang Tuesday May 05, 2020 at 02:45 GMT
Hey Varun,
Sorry to hear that. So we have more info, can you share your config file? Also, has the drop been consistent across random seeds before and after the version change?
Wednesday Apr 29, 2020 at 06:22 GMT
Originally opened as nyu-mll/jiant#1081
Hi,
I'm trying out the MultiRC dataset task(Khasabi et. al 2018). Using the two jiant versions gave the following F1 scores(using BERT-base):
jiant v1.3 = ~58
jiant v1.0 = ~65
with both having the same configurations. I looked over the changes and found one:
v1.0 gives (para + ques + ans option) as BERT input
v1.3 gives ([CLS] para [SEP] ques + ans option [SEP]) as BERT input
I tried doing the same thing as v1.0 in v1.3 but observed no changes in the F1 score(rather F1 decreased).
Kindly let me know if I may try out some other changes between the two versions that I have missed earlier. This drastic drop in performance over version upgrade is hard to understand.
The text was updated successfully, but these errors were encountered: