-
Notifications
You must be signed in to change notification settings - Fork 464
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add japanese bert pretrained model #118
base: master
Are you sure you want to change the base?
Conversation
hi guys, Can you give me the ways to get new BERT pre-trained model ? Thank you. |
Hello, thanks for your contribution. I notice that you didn't change some functions designed to preprocess the english datasets in the data_builder file, and you use the multilingual model to substitute the old one, so I guess that you use the english datasets to train your Japanese model. Is my guess correct? Look forward to your soonest reply. Thank you! |
@congdoanit98 @beanandrew
You're correct. But, I used Japanese dataset. I think these were all changes. Thanks. |
Thanks for your reply.
I am still on the way to change the code, and I haven't do experiments on this, so I wanna know how you solve problems like this, or after your experiments, these problems won't change the experiment results? |
Well, You are correct again.. I conducted my experiments by using the original step5 codes, and I could generate some summaries, and it showed reasonable result. I hope this information will support your work. |
Thanks for your reply, your answer helps me lot!
When I try to debug the code in the Step5, I notice that, because of this code, the labels like [CLS] and [SEP] are skipped, and that cause the Error "CUDAType Error" when I use the dataset it preprocessed. Thus, I changed the code as followed, and the Error no longer appear.
I want to know, do you come with Errors like this when you use the original code? Or is this just my personal problem? |
Hmm, I couldn't find that code you mentioned on master branch. Anyways, your suggestion seems to work beautifully. I remember that I commented out the code below (rouge score calculator), because I couldn't solve errors from pyrouge library. Instead of the pyrouge library, I used this one. I fixed the code to export the summarized text, and after all results were written, I evaluated the rouge score. If you could correctly install the pyrouge library, you don't have to care I think. Additionally, the token size is very important. In Japanese token.txt, I couldn't find something like |
Thank you very much for your quick reply! |
in order to use
bert-base-japanese-whole-word-masking
model, installed transformer independently, and fixed few codes.