-
-
Notifications
You must be signed in to change notification settings - Fork 444
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Adds real-time TTS with Piper 2.0 #21008
Conversation
ok I fucking burst out laughing when I heard Poly, god fucking damnit |
I've taken a look at the projects cited & their datasets, and I must thank & congratulate you for (seemingly) managing to only use A.I. trained on ethically collected data. The only issues I have are that two (US-amy and US-danny) link to https://github.com/MycroftAI/mimic3-voices without saying exactly which of the entries on the page they are (meaning i can't check how the data was collected), and that one (US-kusal) links to https://github.com/MycroftAI/mimic2 , which is not a voice repository or dataset at all but a model for training A.I.s off them. |
Voices amy, danny, kusal retrieved from: Licensing information was not changed when porting files All files demand amy & danny reference MycroftAI/mimic3-voices, which is licensed under Creative Commons Attribution Share Alike 4.0 International kusal references MycroftAI/mimic2, which is licensed under Apache License 2.0 It should be noted that amy and kusal are referenced in the MycroftAI/mimic2 repository, but no differing licenses or datasets are indicated Because the datasets used to train these voices are not available on the repositories linked, it can only be assumed that the repository LICENSE file must be used because of the instruction CC BY-SA 4.0 DEED is complied with by the links provided on the repository README and supplied unadulterated MODEL_CARD files |
I never meant to imply the license was the problem - that's all clearly above-board. I was only concerned with looking through the datasets to see how the data was collected, and my issue was just that not all of them were wholly transparent with where they got their data - I was trying to make sure it was all either from the public domain or collected from people who volunteered to train A.I.. |
I like em |
Maybe some other time, nice experiment. |
Duplicate of #20977 for technical reasons
Document the changes in your pull request
At present, there are 12 voices:
There are several filters that get applied in different ways. Robots and machines will have a silicon filter applied. Lizards, aliens, and ethereals have filters attached to their tongues. Gas masks and sechailers have their own filters as well. Radio will sound more radioy than talking to someone in person.
Once I figure out model training (after this is merged preferably) we can start adding user-provided voices and maybe species-specific voices.
Backend info at https://github.com/yogstation13/yogs-tts
Fix ghosts hearing non-radio TTS from other z levelsWould break multi-z TTSWhy is this good for the game?
Muh immersion
Testing
tts_uhoh.mp4
tts_cat_poly.mp4
Changelog
🆑
rscadd: Added TTS
/:cl: