You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
In principle yes. This would not be much different than the fine-tuning usually done with BERT.
Currently bert-for-tf's does not include the final dense classification layers used in the MLM pre-training task (so they have to be added explicitly in the Keras Model). And if you want to reuse the original weights for those classificationlayers you might have to manually load the weights or carefully name them (by using the exact same names used in the pre-trained checkpoint).
But, I guess, it should be better, if I add a config parameter controlling the instantiation of the cls/ layers, so that the weights from the pre-trained checkpoint are loaded automatically.
Good point, Thank you, @peregilk!
No description provided.
The text was updated successfully, but these errors were encountered: