From b6e109a912b67cc0e9c2e36912d2217c66a4635c Mon Sep 17 00:00:00 2001 From: Symon <59005260+4SAnalyticsnModelling@users.noreply.github.com> Date: Wed, 21 Jun 2023 11:52:01 -0600 Subject: [PATCH] Update Optimisers.jl Add a new variant of Adam optimizer called "PAdam" --- src/Optimisers.jl | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/Optimisers.jl b/src/Optimisers.jl index 20fc8aad..da5df27d 100644 --- a/src/Optimisers.jl +++ b/src/Optimisers.jl @@ -13,7 +13,7 @@ export destructure include("rules.jl") export Descent, Adam, Momentum, Nesterov, Rprop, RMSProp, - AdaGrad, AdaMax, AdaDelta, AMSGrad, NAdam, AdamW, RAdam, OAdam, AdaBelief, + AdaGrad, AdaMax, AdaDelta, AMSGrad, NAdam, AdamW, RAdam, OAdam, AdaBelief, PAdam, WeightDecay, ClipGrad, ClipNorm, OptimiserChain, Lion, AccumGrad