Publication Type
Journal Article
Version
acceptedVersion
Publication Date
3-2024
Abstract
AdamW modifies Adam by adding a decoupled weight decay to decay network weights per training iteration. For adaptive algorithms, this decoupled weight decay does not affect specific optimization steps, and differs from the widely used ℓ2-regularizer which changes optimization steps via changing the first- and second-order gradient moments. Despite its great practical success, for AdamW, its convergence behavior and generalization improvement over Adam and ℓ2-regularized Adam (ℓ2-Adam) remain absent yet. To solve this issue, we prove the convergence of AdamW and justify its generalization advantages over Adam and ℓ2-Adam. Specifically, AdamW provably converges but minimizes a dynamically regularized loss that combines vanilla loss and a dynamical regularization induced by decoupled weight decay, thus yielding different behaviors with Adam and ℓ2-Adam. Moreover, on both general nonconvex problems and PŁ-conditioned problems, we establish stochastic gradient complexity of AdamW to find a stationary point. Such complexity is also applicable to Adam and ℓ2-Adam, and improves their previously known complexity, especially for over-parametrized networks. Besides, we prove that AdamW enjoys smaller generalization errors than Adam and ℓ2-Adam from the Bayesian posterior aspect. This result, for the first time, explicitly reveals the benefits of decoupled weight decay in AdamW. Experimental results validate our theory.
Keywords
Analysis of AdamW, Convergence of AdamW, Generalization of AdamW, Adaptive gradient algorithms
Discipline
Graphics and Human Computer Interfaces
Research Areas
Intelligent Systems and Optimization
Areas of Excellence
Digital transformation
Publication
IEEE Transactions on Pattern Analysis and Machine Intelligence
First Page
1
Last Page
8
ISSN
0162-8828
Identifier
10.1109/TPAMI.2024.3382294
Publisher
Institute of Electrical and Electronics Engineers
Citation
ZHOU, Pan; XIE, Xingyu; LIN, Zhouchen; and YAN, Shuicheng.
Towards understanding convergence and generalization of AdamW. (2024). IEEE Transactions on Pattern Analysis and Machine Intelligence. 1-8.
Available at: https://ink.library.smu.edu.sg/sis_research/8986
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1109/TPAMI.2024.3382294