-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New model editions (GPT4) #340
Comments
Hi @deep-diver |
I think so. You need to trim to get after Response |
But I think the format and prompt template of these two pieces of data are the same. Do you have any understanding of why there is such a difference? |
Hello. |
It looks like the loss not only applies to model genereated output but also such template such as "instruction:" and "input:{input}" |
Sounds reasonable, do you have any attempts to correct it? |
The only way that comes to my mind is to re fine-tune the model and set labels of "instruction, input, etc" to -100. |
Hi @tloen
I have trained the following models on GPT4 generated Alpaca dataset(from the one in this repo), and they are available through Hugging Face Model hub.
You can also find out the link for the training logs on each Model repository.
I hope this might be useful for someone, and I also hope these could be included in the list in this repo.
The text was updated successfully, but these errors were encountered: