Details
-
Improvement
-
Status: Open
-
Major
-
Resolution: Unresolved
-
None
-
None
-
None
Description
The common implementations of adaptive gradient algorithms, such
as Adam, limit the potential benefit of weight decay regularization, because the
weights do not decay multiplicatively (as would be expected for standard weight
decay) but by an additive constant factor.
This following paper found a way to fix regularization in Adam Optimization with one addition step(+ wx) to the gradient step :
https://arxiv.org/pdf/1711.05101.pdf