-
Notifications
You must be signed in to change notification settings - Fork 121
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
KeyError: 'ja' when I use a Japanese wav file to finetune #15
Comments
Hi this is a common problem with the TTS lib coqui-ai/TTS#3384 (comment) , but I was able to adapt it and now in my custom xtts_finetune_webui it is possible to train Japanese finetune quality jp_finetune.mp4 |
Thanks, your new codes work. I got one finetune (with a few audio data) completed until the end inference test. Afterwards, I made another finetune with more training data. This time it get past Epoch 5/6, then there's a PermissionError: [WinError 5] Access is denied . If I run the .bat file in administration, the cmd window will report a missing file in system32 and stops. Edit: I found that deleting the finetune_model/run folder and running step 2 again works. So this may happen when we make a second finetune and the program can’t delete a file in this folder. Traceback (most recent call last): |
My wav file is converted into mono, 22050Hz, 16bit pcm beforehand. I got this error log:
Existing language matches target language
Loading Whisper Model!
Discarding ID3 tags because more suitable tags were found.
Traceback (most recent call last):
File "D:\Long\AI\Audio\xtts-webui\xtts_finetune_webui.py", line 246, in preprocess_dataset
train_meta, eval_meta, audio_total_size = format_audio_list(audio_path, whisper_model = whisper_model, target_language=language, out_path=out_path, gradio_progress=progress)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\scripts\utils\formatter.py", line 160, in format_audio_list
sentence = multilingual_cleaners(sentence, target_language)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\venv\Lib\site-packages\TTS\tts\layers\xtts\tokenizer.py", line 558, in multilingual_cleaners
text = expand_numbers_multilingual(text, lang)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\venv\Lib\site-packages\TTS\tts\layers\xtts\tokenizer.py", line 538, in expand_numbers_multilingual
text = re.sub(_ordinal_re[lang], lambda m: _expand_ordinal(m, lang), text)
~~~~~~~~~~~^^^^^^
KeyError: 'ja'
I got the same error at local and Colab, so maybe something is wrong with Japanese settings?
The text was updated successfully, but these errors were encountered: