Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RoBERTa trained on NSP task #2

Open
prajjwal1 opened this issue Oct 20, 2020 · 0 comments
Open

RoBERTa trained on NSP task #2

prajjwal1 opened this issue Oct 20, 2020 · 0 comments

Comments

@prajjwal1
Copy link

prajjwal1 commented Oct 20, 2020

Hi,

Firstly, thanks for the paper, lots of interesting insights, and for sharing the code. I wanted to try out how RoBERTa would perform when it is coupled with NSP. I know in the original paper, NSP was discarded, but wanted to see how NSP would work out for my use-case.
Did you try with RoBERTa NSP by any chance ? I see that the code defaults to masked LM. I'm not able to find any RoBERTa checkpoint trained on NSP objective on HF model-hub or fairseq repo. If not, I'd have to see how can I perform pre-training myself. Can you please share instructions on pre-training on Bookcorpus and wikipedia ? I see that the paper mentions it, so it will be helpful if you can share it.

Thanks again for the contribution.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant