diff --git a/README.md b/README.md index 353db7cf08..935627e588 100644 --- a/README.md +++ b/README.md @@ -205,7 +205,7 @@ device = "cuda" if torch.cuda.is_available() else "cpu" print(TTS().list_models()) # Init TTS -tts = TTS("tts_models/multilingual/multi-dataset/xtts_v1").to(device) +tts = TTS("tts_models/multilingual/multi-dataset/xtts_v2").to(device) # Run TTS # ❗ Since this model is multi-lingual voice cloning model, we must set the target speaker_wav and language diff --git a/docs/source/inference.md b/docs/source/inference.md index b40445aed6..611a2445bf 100644 --- a/docs/source/inference.md +++ b/docs/source/inference.md @@ -124,7 +124,7 @@ device = "cuda" if torch.cuda.is_available() else "cpu" print(TTS().list_models()) # Init TTS -tts = TTS("tts_models/multilingual/multi-dataset/xtts_v1").to(device) +tts = TTS("tts_models/multilingual/multi-dataset/xtts_v2").to(device) # Run TTS # ❗ Since this model is multi-lingual voice cloning model, we must set the target speaker_wav and language @@ -231,4 +231,4 @@ api.tts_with_vc_to_file( speaker_wav="target/speaker.wav", file_path="ouptut.wav" ) -``` \ No newline at end of file +``` diff --git a/tests/zoo_tests/test_models.py b/tests/zoo_tests/test_models.py index 79aef5cb14..d1c6b67c39 100644 --- a/tests/zoo_tests/test_models.py +++ b/tests/zoo_tests/test_models.py @@ -14,7 +14,6 @@ MODELS_WITH_SEP_TESTS = [ "tts_models/multilingual/multi-dataset/bark", "tts_models/en/multi-dataset/tortoise-v2", - "tts_models/multilingual/multi-dataset/xtts_v1", "tts_models/multilingual/multi-dataset/xtts_v1.1", "tts_models/multilingual/multi-dataset/xtts_v2", ] @@ -83,14 +82,14 @@ def test_xtts(): if use_gpu: run_cli( "yes | " - f"tts --model_name tts_models/multilingual/multi-dataset/xtts_v1 " + f"tts --model_name tts_models/multilingual/multi-dataset/xtts_v1.1 " f'--text "This is an example." --out_path "{output_path}" --progress_bar False --use_cuda True ' f'--speaker_wav "{speaker_wav}" --language_idx "en"' ) else: run_cli( "yes | " - f"tts --model_name tts_models/multilingual/multi-dataset/xtts_v1 " + f"tts --model_name tts_models/multilingual/multi-dataset/xtts_v1.1 " f'--text "This is an example." --out_path "{output_path}" --progress_bar False ' f'--speaker_wav "{speaker_wav}" --language_idx "en"' ) @@ -104,7 +103,7 @@ def test_xtts_streaming(): speaker_wav = [os.path.join(get_tests_data_path(), "ljspeech", "wavs", "LJ001-0001.wav")] speaker_wav_2 = os.path.join(get_tests_data_path(), "ljspeech", "wavs", "LJ001-0002.wav") speaker_wav.append(speaker_wav_2) - model_path = os.path.join(get_user_data_dir("tts"), "tts_models--multilingual--multi-dataset--xtts_v1") + model_path = os.path.join(get_user_data_dir("tts"), "tts_models--multilingual--multi-dataset--xtts_v1.1") config = XttsConfig() config.load_json(os.path.join(model_path, "config.json")) model = Xtts.init_from_config(config)