Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some questions about training model #862

Closed
guoaoo opened this issue Oct 6, 2021 · 1 comment
Closed

Some questions about training model #862

guoaoo opened this issue Oct 6, 2021 · 1 comment

Comments

@guoaoo
Copy link

guoaoo commented Oct 6, 2021

I have trained a synthesizer used by default hparams from scratch in 14 days, and there are some plot in Tensorboard.

tts_schedule = [(2, 1e-3, 20_000, 12), # Progressive training schedule
(2, 5e-4, 40_000, 12), # (r, lr, step, batch_size)
(2, 2e-4, 80_000, 12), #
(2, 1e-4, 160_000, 12), # r = reduction factor (# of mel frames
(2, 3e-5, 320_000, 12), # synthesized for each decoder iteration)
(2, 1e-5, 640_000, 12)], # lr = learning rate

image
image

Some questions to ask:

  1. The pretrain model is trained to 295k, is that because it is good enough? The size of the .pt file does not increase with the number of training steps.
  2. I used the default hparams during training and only used about 6G of GPU memory, there is still a lot left for 12G 2080ti and 32G V100, how do I adjust the hparams to speed up the training? If increase batch_size, how much should I increase the learning rate? Is reduction factor an adjustable number, and will it have a bad effect on the results?
  3. After reading Training a new model based on LibriTTS #449 , do you recommend that I modify the hparams and train a synthesizer from scratch or just fine-tuning on pretained model?
@ghost
Copy link

ghost commented Oct 6, 2021

  1. No, the number of steps is arbitrary though I believe the training had converged earlier than that.
  2. You need to experiment with batch_size and learning rate to find the best values. The reduction factor r is adjustable (integer >= 1). Training and inference are faster when it is set to a higher number. Quality is better when r is lower.
  3. Train from scratch.

@ghost ghost closed this as completed Oct 6, 2021
This issue was closed.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant