You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I can't find in the repository the code used to continue mbart pretraining to create mbarthez. Did you make it available somewhere ?
More specifically, I'm interested in understanding how you adapted the mbart tokenizer. It looks like that the checkpoint on huggingface uses the barthez tokenizer, not the mbart tokenizer. So my question is: how did you align the pretrained mbart embeddings with the barthez tokenizer vocab ?
The text was updated successfully, but these errors were encountered:
Hello @moussaKam ,
I can't find in the repository the code used to continue mbart pretraining to create mbarthez. Did you make it available somewhere ?
More specifically, I'm interested in understanding how you adapted the mbart tokenizer. It looks like that the checkpoint on huggingface uses the barthez tokenizer, not the mbart tokenizer. So my question is: how did you align the pretrained mbart embeddings with the barthez tokenizer vocab ?
The text was updated successfully, but these errors were encountered: