-
Notifications
You must be signed in to change notification settings - Fork 4
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
See how replacing random weights with pretrained and fine-tuned weights in MATCH affects performance #72
Comments
According to @elkong the weights were empty for the original authors. This probably didn't effect them that much since their dataset was large. For us, I think we need to take a bert or GPL model, expand pretrain data with our dataset and new words so that it understand the relationship between these words. Then! we surgically add it to match. |
This discussion I had with the author might be relevant to this: On Friday, June 11, 2021 at 11:26 AM Yu Zhang wrote:
On Fri, Jun 11, 2021 at 8:07 AM Ruffridge, Brandon wrote:
|
related to #24 |
|
"one of my other major bottlenecks is pretraining weights – I’ve been training MATCH from random weight initializations every time, whereas with models like GPT-2 people just take the pretrained weights and finetune them to get state-of-the-art results. So I’ll look into either finding a way to start from pretrained MATCH weights, or finetuning GPT-2 or some such model." - Eric
The text was updated successfully, but these errors were encountered: