Skip to content

Commit

Permalink
[Benchmark] Enable use_fast_layer_norm for llama2 benchmark (#8714)
Browse files Browse the repository at this point in the history
  • Loading branch information
deepllz authored Jul 4, 2024
1 parent fd01043 commit 70564ba
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 0 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
"sequence_parallel": 0,
"use_flash_attention": true,
"use_fused_rms_norm": true,
"use_fast_layer_norm": true,
"fuse_attention_ffn": true,
"fuse_attention_qkv": true,
"use_fused_rope": true,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
"sequence_parallel": 1,
"use_flash_attention": true,
"use_fused_rms_norm": true,
"use_fast_layer_norm": true,
"fuse_attention_ffn": true,
"fuse_attention_qkv": true,
"use_fused_rope": true,
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -16,6 +16,7 @@
"sequence_parallel": 0,
"use_flash_attention": true,
"use_fused_rms_norm": true,
"use_fast_layer_norm": true,
"fuse_attention_ffn": true,
"fuse_attention_qkv": true,
"use_fused_rope": true,
Expand Down

0 comments on commit 70564ba

Please sign in to comment.