Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add llama-2 template support for fine-tuning #2423

Closed

Conversation

karthik19967829
Copy link

Make llama-2 template as default to support fine-tuning llama-2 models better

Why are these changes needed?

This changes changes the template in train.py to llama-2 . From my experiments I reached better convergence and starting loss with llama-2 template for the llama-2 7B model than the vicuna template . Considering a significant part of the community is using llama-2 it might be good to make it the default template , or have a mechanism to support llama-2

Related issue number (if applicable)

Fixes #2043

Checks

  • I've run format.sh to lint the changes in this PR.
  • I've included any doc changes needed.
  • I've made sure the relevant tests are passing (if applicable).

Make llama-2 template as default to support fine-tuning llama-2 models better
@merrymercy
Copy link
Member

  1. Can we make this an argument conversation_template in
    class TrainingArguments(transformers.TrainingArguments):
    ? Let us keep the vicuna as the default one and gradually move to llama-2 as the default
  2. Could you fix the format?

@karthik19967829
Copy link
Author

Sure will make these changes

@merrymercy
Copy link
Member

closed due to inactivity

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

Fine tune llama 2
2 participants