Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Finetuning trained model #75

Closed
aleksas opened this issue May 19, 2018 · 3 comments
Closed

Finetuning trained model #75

aleksas opened this issue May 19, 2018 · 3 comments

Comments

@aleksas
Copy link
Contributor

aleksas commented May 19, 2018

In reference to recommendation how should repeated training be executed for the second time?
Should checkpoint path point to checkppoints/checkpoint_step...._ema.pth or not ema checkpoint? And what does ema stands for?

@r9y9
Copy link
Owner

r9y9 commented May 19, 2018

EMA means Exponential Moving Average (EMA). See details for Tacotron 2 paper: https://arxiv.org/abs/1712.05884. There's no clear answer which to use for finetuning, but I usually used EMA version of the checkpoint when I trained a model for sufficient time (e.g, over 2 days for MoL case).

I'll leave some commands I used for LJSpeech experiments which might help as follows:

  • Finetuning trained model
python train.py --data-root=./data/ljspeech/ \
  --checkpoint-dir=./checkpoints_mixture \
  --log-event-path=./log/wavenet_mixture \
  --restore-parts=./pretrained_models/20180127_mixture_lj_checkpoint_step000410000_ema.pth
  • Accidentally training aborted (my PC was rebooted), then resuming
python train.py --data-root=./data/ljspeech/ \
  --checkpoint-dir=./checkpoints_mixture \
  --log-event-path=./log/wavenet_mixture \
  --checkpoint=./checkpoints_mixture/checkpoint_step000256701.pth

The two commands above were actually used for training the pre-traind model.

@aleksas
Copy link
Contributor Author

aleksas commented May 19, 2018

Thanks! Will try to fine tune my model now.

@d2sys
Copy link

d2sys commented Aug 8, 2018

Anyone can share the minimum number of iterations which is enough for fine-tuning?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants