Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Is it possible to use this library for doing additional domain specific pre-training of the Bert weights? #13

Open
peregilk opened this issue Oct 23, 2019 · 1 comment
Assignees

Comments

@peregilk
Copy link

No description provided.

@kpe
Copy link
Owner

kpe commented Oct 23, 2019

In principle yes. This would not be much different than the fine-tuning usually done with BERT.
Currently bert-for-tf's does not include the final dense classification layers used in the MLM pre-training task (so they have to be added explicitly in the Keras Model). And if you want to reuse the original weights for those classification layers you might have to manually load the weights or carefully name them (by using the exact same names used in the pre-trained checkpoint).
But, I guess, it should be better, if I add a config parameter controlling the instantiation of the cls/ layers, so that the weights from the pre-trained checkpoint are loaded automatically.
Good point, Thank you, @peregilk!

@kpe kpe self-assigned this Oct 23, 2019
@kpe kpe mentioned this issue Jan 6, 2020
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants