Trained model 600 epochs complete flop #310
Replies: 3 comments 3 replies
-
I think I started too big, im going to go over dataset manually and get best samples 3-8 seconds and feed it 20 minutes, see what it does |
Beta Was this translation helpful? Give feedback.
-
So after feeding it 20 minutes of selected audio about 300 segments , I got better results but still not a predictable voice tonation or clear voice |
Beta Was this translation helpful? Give feedback.
-
Hi @Eyalm321 You are best reading up on the Coqui forums about training, however, 600 epochs is a lot of training. That's approaching the kind of training you would use to train a new language. If you are just training a voice in a language that the model supports e.g. English, you wont need a huge amount of epochs (the defaults should work). You can over-train a model. Its best to start low and you can always further train a model if needs be. There is no actual perfect recipe for training a model however and ask Coqui note https://docs.coqui.ai/en/latest/faq.html#how-do-i-know-when-to-stop-training The absolute most important thing is the quality of the training data and dataset. Higher quality audio and ensuring the eval and train csv files match the spoken audio, gives the best results. |
Beta Was this translation helpful? Give feedback.
-
Hey guys,
I need your help with some tips for training
I fed the model around hour and half of single speaker voice about 2200 segments 1-5 seconds average
Did I over segment it with pyannote?
Should I fed it more data? Less?
How many epochs is average to get good result? Whats your process?
Seems like after 600 epochs its talking complete giberish
Beta Was this translation helpful? Give feedback.
All reactions