-
Notifications
You must be signed in to change notification settings - Fork 27.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TF mT5 model is not adding new tokens into it's vocabulary. #13839
Comments
This seems to be an issue with TF model cc @Rocketknight1 @patrickvonplaten it seems there are no extra tokens in mt5 tokenizer and there's a mismatch between |
Actually, everything looks fine for me here... TFMT5Model and MT5Tokenier have a tokenizer mismatch because of the same reason as T5, see: #4875 Apart from this the following code works: from transformers import TFMT5ForConditionalGeneration, T5Tokenizer
model = TFMT5ForConditionalGeneration.from_pretrained("google/mt5-base")
tokenizer = T5Tokenizer.from_pretrained("google/mt5-base")
tokenizer.add_special_tokens({'bos_token':'','eos_token':''})
model.resize_token_embeddings(len(tokenizer)) @laddhakrishna - is there any reason you used |
Sir even if we use --> model.resize_token_embeddings(...) ,still we are getting the same error. |
Hey @laddhakrishna, could you provide a google colab that reproduces the error with |
Colab notebook link: Please have a look at it sir. |
Hey @laddhakrishna, Thanks for the colab I can reproduce! |
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread. Please note that issues that do not follow the contributing guidelines are likely to be ignored. |
Maybe @Rocketknight1 or @gante can take a look? |
Environment info
transformers
version: 4.11.2Who can help @patrickvonplaten, @patil-suraj
Information
Model I am using (Bert, XLNet ...):
I am using mT5 model. The problem arised when I am trying to resize the token embeddings. I added new special tokens to my tokenizer's vocabulary, when I resized the model's token embeddings according to the length of the tokenizer, I am facing the issue.
Link for the notebook: https://colab.research.google.com/drive/1ooKa5aQ_FAEnicxBL8bJnNAO4H9vFuv-?usp=sharing
Code from the notebook:
!pip install transformers
!pip install sentencepiece
from transformers import TFMT5ForConditionalGeneration, T5Tokenizer
model = TFMT5ForConditionalGeneration.from_pretrained("google/mt5-base")
tokenizer = T5Tokenizer.from_pretrained("google/mt5-base")
tokenizer.add_special_tokens({'bos_token':'','eos_token':''})
model._resize_token_embeddings(len(tokenizer))
The error message is:
ValueError: Attempt to convert a value (None) with an unsupported type (<class 'NoneType'>) to a Tensor.
The problem arises when using:
The tasks I am working on is:
To reproduce
Steps to reproduce the behavior:
Expected behavior
The text was updated successfully, but these errors were encountered: