-
-
Notifications
You must be signed in to change notification settings - Fork 130
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hindi Finetuning Not supported yet? #424
Comments
@Mark2619 Have posted an update, so you will need to You would have to use an XTTS 2.0.3 model, the earlier ones dont support Hindi. The tokenizers were only extended now to fully support Hindi training 1x month ago. I have not tested, so if it goes wrong/doesn't work, your best bet would be to go here https://github.com/idiap/coqui-ai-TTS/issues where they work keeping the back end Coqui scripts alive and working. Thanks |
Its definitely there This is the commit: af29b1e You can always manually download from here https://github.com/erew123/alltalk_tts/blob/alltalkbeta/finetune.py down arrow two over from the word "raw" at the top right |
Oh yes manually downloading did work, but i'm getting some error." To create a public link, set |
Thanks for letting me know. As mentioned I would suggest you address this up the chain to the people whom maintain the Coqui TTS engine/scripts. https://github.com/idiap/coqui-ai-TTS/issues. Long and short of it is, this means that the Coqii training scripts don't know what "hi" the language code for Hindi is If I get time in future I will look further into it, however as I say, they manage the core scripts so I have no control over that and Hindi support on XTTS was a pretty much undocumented feature by Coqui. |
Thanks, I'll take it up with coqui to further fix this issue. Just a quick question, if i process the data manually (split up audio and create metadata) the training part would still work fine, right? because right now the only issue I see is dataset creation. |
If it happened at that stage then thats Whisper. Which whisper model did you use? https://github.com/openai/whisper?tab=readme-ov-file#available-models-and-languages |
large-v3 is the one i used |
It should have worked in theory. OpenAI's code accepts And Large-V3 is multi-language. There is nothing special about the AllTalk whisper installation, so its whatever OpenAI are supporting should work. Yes you can build a dataset yourself. Using it is documented in the interface of finetuning (as I recall). The Coqui Trainer structure for the CSV files is detailed here https://docs.coqui.ai/en/latest/formatting_your_dataset.html The simplest way to understand it though, would be to quickly do step 1 as an English or other language set which will build the files for you to see how they are laid out on disk and within the CSV files. Id suggest you make a small Hindi dataset after that for testing purposes. You will need to change the language in the lang.txt file to |
Thanks again for the detailed info, I will try to train on a small dataset and update here for anyone else looking to see if it works or not. Cheers! |
Hi, just wanted to finetune a model and i see there is no Hindi language in the finetune training tab. Is it not available yet?
The text was updated successfully, but these errors were encountered: