Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Test of analysis on open-domain QA #495

Open
2 of 5 tasks
neubig opened this issue Sep 18, 2022 · 1 comment
Open
2 of 5 tasks

Test of analysis on open-domain QA #495

neubig opened this issue Sep 18, 2022 · 1 comment

Comments

@neubig
Copy link
Contributor

neubig commented Sep 18, 2022

There is a nice PrimeQA framework developed by IBM, and it seems that it should make it relatively easy to generate open-domain QA results from multiple SOTA models.

It would be nice if we could look into these results, analyze them with explainaboard, and see if there are any ways we could use them to improve our analysis of Open Domain QA models.

In order to do this, we'd need to:

  • Take a look at the machine reading comprehension tutorial for primeqa
  • Decide which datasets we want to focus on
  • Generate multiple system outputs for these datasets
  • Analyze them in ExplainaBoard and see if we get any interesting insights
  • Further add features to the analysis
@neubig
Copy link
Contributor Author

neubig commented Oct 14, 2022

Suggestion from Avi at IBM, so I guess we should focus on PrimeQA's TyDiQA:

We’ve been focusing on TyDI for sometime now: the model’s available here too: https://huggingface.co/PrimeQA/tydiqa-primary-task-xlm-roberta-large. So if you want to do the same expts with TyDI and use PrimeQA you can totally get the system outputs by running decoding. It should be pretty straightforward.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant