-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Multilingual Tuned Model Translating everything to "sssssssss" #79
Comments
That is weird. Maybe fine-tuning ran for too long on very smallish data sets and the model heavily overfitted to the fine-tuning data set and forgot everything else? Did you see strange perplexity scores during fine-tuning? |
It looked like it only went through one round? What is even weirder is the compare file (Tatoeba-test-v2021-08-07.afr-eng.opus-tuned4afr2eng.spm1k-spm1k1.transformer-align.afr.eng)
And then of course the eval file records the bleu score as 0. I double checked all the data...I used about 1500 lines of afr-eng data to finetune mul-eng model). I am really at a loss here because I can tune monolingual models just fine using the same steps. Do you have anymore insight? |
I was able to successfully tune a multilingual model utilizing data for one of the included languages. Unfortunately, when the translations are run through the model, they all result in "ssssss" (even if a blank line is sent). I can't find anything about this happening with Marian models after tuning and can't seem to figure out the issue. The source and target data files all look fine, & there are no errors in the tuning process. Have you ever seen this happen at all, or have any idea on what it could be? I am really stuck. Thank you!
The text was updated successfully, but these errors were encountered: