Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Multispeaker trained model inferencing different voices #260

Open
BakuDev opened this issue Dec 16, 2021 · 1 comment
Open

Multispeaker trained model inferencing different voices #260

BakuDev opened this issue Dec 16, 2021 · 1 comment

Comments

@BakuDev
Copy link

BakuDev commented Dec 16, 2021

Trained a model with dataset of a multiple speakers.
Quality is ok but... The model produces random speaker voice on inference.
If there any type of control on this, is it possible to choose the voice?
How the model chooses which one to use for inference?
What's interesting - model picks the same voice for each specific text (unless I edit anything in it, even a dot or comma).

@hongchengzhu
Copy link

Trained a model with dataset of a multiple speakers. Quality is ok but... The model produces random speaker voice on inference. If there any type of control on this, is it possible to choose the voice? How the model chooses which one to use for inference? What's interesting - model picks the same voice for each specific text (unless I edit anything in it, even a dot or comma).

Hello, have you solved this issue? Recently I'd like to use waveglow as my vocoder in the multispeaker setting (VCTK Corpus).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants