Skip to content

Commit 9fd8242

Browse files
authored
Update meta-llama-Llama-2-7b_pretrain_dynamic_auto_bs32_bf16_Sharding32_Stage2.sh
1 parent ddcb722 commit 9fd8242

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

tests/test_tipc/static/auto_parallel/llama2/N4C32/meta-llama-Llama-2-7b_pretrain_dynamic_auto_bs32_bf16_Sharding32_Stage2.sh

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,10 @@ param+='dynamic_auto=_dynamic_auto '
2323
# This optimization currently only runs in the dynamic automatic parallelism of Llama7B.
2424
export FLAGS_fuse_reducescatter_in_opt=1
2525

26+
# Enable tensor fusion and sharding overlap optimization
27+
export FLAGS_enable_tensor_fusion=1
28+
export FLAGS_enable_sharding_overlap=1
29+
2630
cd ./tests
2731
bash ./test_tipc/static/auto_parallel/llama2/benchmark_common/prepare.sh
2832

0 commit comments

Comments
 (0)