Title
NAMSG: An Efficient Method For Training Neural Networks.
Abstract
We introduce NAMSG, an adaptive first-order algorithm for training neural networks. The method is efficient in computation and memory, and straightforward to implement. It computes the gradients at configurable remote observation points, in order to expedite the convergence by adjusting the step size for directions with different curvatures, in the stochastic setting. It also scales the updating vector elementwise by a nonincreasing preconditioner, to take the advantages of AMSGRAD. We analyze the convergence properties for both convex and nonconvex problems, by modeling the training process as a dynamic system, and provide a guideline to select the observation distance without grid search. We also propose a datadependent regret bound, which guarantees the convergence in the convex setting. Experiments demonstrate that NAMSG works well in practice and compares favorably to popular adaptive methods, such as ADAM, NADAM, and AMSGRAD.
Year
Venue
DocType
2019
arXiv: Learning
Journal
Volume
Citations 
PageRank 
abs/1905.01422
0
0.34
References 
Authors
0
8
Name
Order
Citations
PageRank
chen yushu112.37
Hao Jing200.68
Wenlai Zhao3176.42
Zhiqiang Liu4124.68
Liang Qiao501.69
Wei Xue640052.95
Haohuan Fu749163.94
Guangwen Yang859992.40