-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bugs when fine-tune tiny-llama with instructions using tiny-llama's conversation template #2992
Comments
Hi, hychaochao could you help to confirm if this pr #2996 fix the issue? |
Yes, it works!!!Thanks for your great work again!
and this is part of the output:
|
The script I used for testing was wrong. I was testing the train.py file instead of the train_with_template.py file. Now I retested and found that it still doesn't work. This is my training arguments:
and this is part of the output:
|
@congchan and I met the same error when I use train_with_template.py to fine-tune the llama-2 by using llama-2's conversation template.
this is part of the output:
|
Hi, @hychaochao I just testing with llama 2 and tinyllama and they both works with this fix: Fee free to confirm the results on your data, and let me know if it works. Thank you. |
@congchan Very happy to see that you have fixed the bug! I have tried it on my data and it works. Thank you again for such a great job and such efficiency! |
Thanks for your suggestions! |
Thanks for your great work! I met some problems when using train_with_template.py to fine-tune the tinyllama by using tinyllama's conversation template.
This is my script:
However, the tokenization mismatch warning was reported, and the training loss was always 0.
just like #2871
I've confirmed that I'm using tinyllama's template when training, but it still doesn't work.
The text was updated successfully, but these errors were encountered: