Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Reproduce results in the paper #10

Closed
a3616001 opened this issue Oct 1, 2019 · 3 comments
Closed

Reproduce results in the paper #10

a3616001 opened this issue Oct 1, 2019 · 3 comments

Comments

@a3616001
Copy link

a3616001 commented Oct 1, 2019

Hi,

I was trying to reproduce results by running your code, and couldn't get exactly the same precision on SQuAD.
Here is what I got for bert_large model on SQuAD:
all_samples: 303
list_of_results: 303
global MRR: 0.3018861233236291
global Precision at 10: 0.5676567656765676
global Precision at 1: 0.16831683168316833

However, in the paper, the table shows that there should be 305 samples and the precision should be 17.4%.

At first, I guessed that it is because 2 samples are excluded because their object labels are out of the common vocabulary, but even after testing without common vocabulary, I got global Precision at 1: 0.1704918, which is still different to results in the paper.

Is there a way to reproduce the same results in the paper?
Please correct me if I made any mistakes! Thanks!

@fabiopetroni
Copy link
Contributor

fabiopetroni commented Oct 2, 2019

Hey @a3616001,

strange.
Just re-executed the run_experiments scripts and I get P@1 : 0.1737704918032787 for the BERT-large model. Are you using BERT-large?
Also, the script should use all the 305 examples.
This is how your output should look like:
Screenshot 2019-10-02 10 47 33

@jeslev
Copy link

jeslev commented Jan 28, 2021

Hi, @a3616001 did you finally get the results from the paper?
I got the same results as you (skipping 2 examples after the filter_samples function).

Thanks in advance

@Hannibal046
Copy link

Same problem

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

5 participants
@jeslev @a3616001 @fabiopetroni @Hannibal046 and others