You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: paddlenlp/trainer/training_args.py
+6-1Lines changed: 6 additions & 1 deletion
Original file line number
Diff line number
Diff line change
@@ -1109,7 +1109,12 @@ class TrainingArguments:
1109
1109
def__post_init__(self):
1110
1110
world_size=paddle.distributed.get_world_size()
1111
1111
ifin_auto_parallel_align_mode():
1112
-
self.max_grad_norm=0.0
1112
+
# self.max_grad_norm = 0.0
1113
+
# The current auto_hybrid_pp has aligned the handling of ClipGradByGlobalNorm with the original dygraph semi-auto parallel and dynamic manual-parallel modes and can correctly handle grad_clip, so it is no longer necessary to set max_grad_norm=0.0.
1114
+
ifself.max_grad_norm!=0.0:
1115
+
warnings.warn(
1116
+
"max_grad_norm is not 0.0,We will execute ClipGradByGlobalNorm,if you want to disable it,please set max_grad_norm=0.0"
0 commit comments