Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why is the model trained with audio channels as 2? Although training data is LJSpeech. #4

Open
Ashigarg123 opened this issue Oct 2, 2024 · 1 comment

Comments

@Ashigarg123
Copy link

Am I missing something?

@signofthefour
Copy link
Collaborator

@Ashigarg123 Hi, Sorry for late replying.
It is my bad for laziness, I reused the named of parameters as in PriorGrad. Here is why audio channels = 2.

First, FreGrad predicts Wavelet features instead of Waveform. Therefore, the output will be [low_freq, high_freq] instead of [waveform]. So that, the number of output channel is 2 (each for low and high separately) instead of 1 (mono-channel waveform).

Please discuss more if it remain ambiguous.

@signofthefour signofthefour pinned this issue Nov 12, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants