Skip to content

Commit

Permalink
[LLM] fix lora target modules on llama
Browse files Browse the repository at this point in the history
  • Loading branch information
SylarTiaNII committed May 7, 2024
1 parent fdcabf8 commit 252fdc4
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion llm/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -125,9 +125,10 @@ def get_lora_target_modules(model):
".*v_proj.*",
".*k_proj.*",
".*o_proj.*",
".*gate_proj.*",
".*qkv_proj.*" ".*gate_proj.*",
".*down_proj.*",
".*up_proj.*",
".*gate_up_fused_proj.*",
]
elif model.base_model_prefix == "opt":
target_modules = [
Expand Down

0 comments on commit 252fdc4

Please sign in to comment.