On Convergence of Adam fo
On Convergence of Adam for Stochastic Optimization under Relaxed Assumptions
On Convergence of Adam for Stochastic Optimization under Relaxed Assumptions
arXiv:2402.03982v2 Announce Type: replace-cross
Abstract: The Adaptive Momentum Estimation (Adam) algorithm is highly effective in training various deep learning tasks. Despite this, there’s limited theoretical understanding for Adam, especially when focusing on its vanilla form in non-convex smoot…