-
Notifications
You must be signed in to change notification settings - Fork 39
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Train sentences #16
Comments
Hi, glad it's been working for you. If there is a lot of ambiguity I would say it could work if you feed the whole sentences, albeit quite memory hungry. You could also try to use the autoregressive model and feed each word plus context as input and just the word-phonemes as target (make sure you use some word separator for the context). |
Hi, Thanks a lot for your work! Relating to this issues, I would like to get some advice on training my model. Specifically, I want to see whether a transformer/autoregressive transformer model could learn liaison in French. To this end, I generated a training data where one or two words are as grapheme and their corresponding phonemic transcription as phoneme listed in each line (e.g., line1: Nous / nu, line2: Nous étions / nuz etjɔ̃, here you can see that liaison /z/ in the word 'nous' occurs in a certain context, in this case /e/ after the first word). I have trained with the two models and for some reason the model failed to transcribe liaison. I have updated text/phoneme symbols in the config file and decreased the batch size to 16. Other than than, all other things remained same as it was. I also checked the number of liaison occurrence to address potential imbalance between cases with liaison and cases without liaison. Do you think I miss something such that the trained model does not manage to detect the context to produce liaison at all? Thanks for your advice in advance. |
Hi. I was able to train an italian model almost perfectly with the exception of few words that are intrinsecally ambiguous without context. Since your model is similar to the bert transformer what do you think would be the best solution to let the model learn word with context? Passing the sentences would be enough? Or a MLM should be implemented?
The text was updated successfully, but these errors were encountered: