Skip to content

Commit 5f74e51

Browse files
committed
run llama13b custom op perf
1 parent 850c6c2 commit 5f74e51

File tree

1 file changed

+6
-5
lines changed

1 file changed

+6
-5
lines changed

tests/test_tipc/static/auto_parallel/llama2/pretrain_config_llama2_13b/pretrain-llama2_13b.json

Lines changed: 6 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -7,14 +7,14 @@
77
"gradient_accumulation_steps": 4,
88
"per_device_eval_batch_size": 4,
99
"tensor_parallel_degree": 1,
10-
"pipeline_parallel_degree": 4,
10+
"pipeline_parallel_degree": 1,
1111
"sharding": "stage1",
1212
"data_parallel_config": "enable_allreduce_avg_in_gradinent_scale gradient_sync_after_accumulate",
1313
"sharding_parallel_config": "enable_overlap enable_tensor_fusion",
1414
"tensor_parallel_config": "enable_mp_async_allreduce",
1515
"pipeline_parallel_config": "enable_send_recv_overlap enable_split_backward",
16-
"pipeline_schedule_mode": "VPP",
17-
"virtual_pp_degree": 5,
16+
"pipeline_schedule_mode": "",
17+
"virtual_pp_degree": 1,
1818
"sequence_parallel": 0,
1919
"use_flash_attention": true,
2020
"use_fused_rms_norm": true,
@@ -51,6 +51,7 @@
5151
"recompute_granularity": "full",
5252
"save_total_limit": 2,
5353
"device": "gpu",
54-
"to_static": true,
55-
"enable_auto_parallel": true
54+
"to_static": false,
55+
"enable_auto_parallel": true,
56+
"num_hidden_layers": 10
5657
}

0 commit comments

Comments
 (0)