-
Notifications
You must be signed in to change notification settings - Fork 2.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training Scripts masked LM #263
Comments
Hi, can you specify which project you're directing your question to? |
Sorry, my question is to the Perceiver IO-project-team. In the NLP-world often the pretrained models are just english or "all the worlds languages". Many users however need inference in non-english languages and have 1 or 2 GPUs rather than TPU-pods, so for them it's most efficient to pretrain only in the language you actually need inference in. So both for pretraining and finetuning it'd be great to have the scripts you used in your pretraining of the masked LM available. Thx |
Hi, thanks for your interest in Perceiver IO. We do not plan on open sourcing the training scripts for the masked LM, because the script is heavily tied to our internal infrastructure for training these models at scale. We do have an example training pipeline for ImageNet released as well as the exact configuration we used for language modeling from bytes (in the language modeling colab), which hopefully would be of use if you wish to train a new language model from scratch for other languages. Do let us know if you have any further questions or if you encounter any issues trying to replicate our work! |
Thx for the orientation. I will then get my head around the ImageNet-pipeline and try to adapt that to the NLP case. |
Hi @fding |
Do you intend to publish the training scrips for the masked LM as well?
The text was updated successfully, but these errors were encountered: