You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
(2, 1e-4, 160_000, 12), # r = reduction factor (# of mel frames
(2, 3e-5, 320_000, 12), # synthesized for each decoder iteration)
(2, 1e-5, 640_000, 12)], # lr = learning rate
Some questions to ask:
The pretrain model is trained to 295k, is that because it is good enough? The size of the .pt file does not increase with the number of training steps.
I used the default hparams during training and only used about 6G of GPU memory, there is still a lot left for 12G 2080ti and 32G V100, how do I adjust the hparams to speed up the training? If increase batch_size, how much should I increase the learning rate? Is reduction factor an adjustable number, and will it have a bad effect on the results?
After reading Training a new model based on LibriTTS #449 , do you recommend that I modify the hparams and train a synthesizer from scratch or just fine-tuning on pretained model?
The text was updated successfully, but these errors were encountered:
No, the number of steps is arbitrary though I believe the training had converged earlier than that.
You need to experiment with batch_size and learning rate to find the best values. The reduction factor r is adjustable (integer >= 1). Training and inference are faster when it is set to a higher number. Quality is better when r is lower.
I have trained a synthesizer used by default hparams from scratch in 14 days, and there are some plot in Tensorboard.
Real-Time-Voice-Cloning/synthesizer/hparams.py
Lines 52 to 57 in 7432046
Some questions to ask:
.pt
file does not increase with the number of training steps.batch_size
, how much should I increase thelearning rate
? Isreduction factor
an adjustable number, and will it have a bad effect on the results?The text was updated successfully, but these errors were encountered: