Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Quick question: where do the pretrained model's phoneme dictionaries come from? #33

Open
ADD-eNavarro opened this issue Mar 24, 2023 · 3 comments

Comments

@ADD-eNavarro
Copy link

Hi!

Great work you're doing here.
I've been testing your tool, it's easy to use and gives fine results.
Since I'm looking for a tool to generate a phonemized imput for the VITS model (in onnx format), I need to use the same tokenizer (phonemizer) that model espects. I've found that your pretrained models already have the dictionary embedded in them. Can I ask where did those dictionaries come from? In your colab training example you use CUNY-CL/wikipron's ones, but I was wondering if those are the ones you used originally or just in the example.

Thanks.

@NextDevX
Copy link

Hi,

Have you found a way to convert the deep phonemizer module to onnx format and use it?

@ADD-eNavarro
Copy link
Author

I'm afraid I didn't. I mean, nobody answered my question, so there was nowhere else for me to go. I just let down VITS.

@NextDevX
Copy link

NextDevX commented May 17, 2024

@ADD-eNavarro

[Is There a Sample Showing How to Convert to ONNX?](#41 (comment))

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants