-
Notifications
You must be signed in to change notification settings - Fork 119
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
KeyError: 'ja' when I use a Japanese wav file to finetune #15
Comments
Hi this is a common problem with the TTS lib coqui-ai/TTS#3384 (comment) , but I was able to adapt it and now in my custom xtts_finetune_webui it is possible to train Japanese finetune quality jp_finetune.mp4 |
Thanks, your new codes work. I got one finetune (with a few audio data) completed until the end inference test. Afterwards, I made another finetune with more training data. This time it get past Epoch 5/6, then there's a PermissionError: [WinError 5] Access is denied . If I run the .bat file in administration, the cmd window will report a missing file in system32 and stops. Edit: I found that deleting the finetune_model/run folder and running step 2 again works. So this may happen when we make a second finetune and the program can’t delete a file in this folder. Traceback (most recent call last): |
My wav file is converted into mono, 22050Hz, 16bit pcm beforehand. I got this error log:
Existing language matches target language
Loading Whisper Model!
Discarding ID3 tags because more suitable tags were found.
Traceback (most recent call last):
File "D:\Long\AI\Audio\xtts-webui\xtts_finetune_webui.py", line 246, in preprocess_dataset
train_meta, eval_meta, audio_total_size = format_audio_list(audio_path, whisper_model = whisper_model, target_language=language, out_path=out_path, gradio_progress=progress)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\scripts\utils\formatter.py", line 160, in format_audio_list
sentence = multilingual_cleaners(sentence, target_language)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\venv\Lib\site-packages\TTS\tts\layers\xtts\tokenizer.py", line 558, in multilingual_cleaners
text = expand_numbers_multilingual(text, lang)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\Long\AI\Audio\xtts-webui\venv\Lib\site-packages\TTS\tts\layers\xtts\tokenizer.py", line 538, in expand_numbers_multilingual
text = re.sub(_ordinal_re[lang], lambda m: _expand_ordinal(m, lang), text)
~~~~~~~~~~~^^^^^^
KeyError: 'ja'
I got the same error at local and Colab, so maybe something is wrong with Japanese settings?
The text was updated successfully, but these errors were encountered: