Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

About the application #1

Open
Lerry123 opened this issue Sep 20, 2020 · 3 comments
Open

About the application #1

Lerry123 opened this issue Sep 20, 2020 · 3 comments

Comments

@Lerry123
Copy link

Hello,I have used the sequnet with residual in the speech enhancement. When used the large data in train and test,the result is so bad.Could you give some suggestion about this issue?

@f90
Copy link
Owner

f90 commented Sep 22, 2020

Hey, can you clarify more about how you applied the Seq-U-Net? I need some more info about the setting so I can give you some clues as to what the problem might be!

@Lerry123
Copy link
Author

I sampled the speech using a sliding window of about 1s, which was about 16,384 sample points. The sample points were input into the SEQ-U-NET used for speech waveform generation in the original text, and the number of channels was changed to 1.
The VCTK database was used for training, the train set is 11578, and the test set is 874. When all the samples were used for training, the speech was distorted.But when I trained with 50 samples and tested with 30, there was no distortion in the enhanced speech.

@Lerry123
Copy link
Author

I have used raw_audio in original code. I wrote the input size and output size as 16384.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants