Title
KALMAN OPTIMIZER FOR CONSISTENT GRADIENT DESCENT
Abstract
Deep neural networks (DNN) are typically optimized using stochastic gradient descent (SGD). However, the estimation of the gradient using stochastic samples tends to be noisy and unreliable, resulting in large gradient variance and bad convergence. In this paper, we propose Kalman Optimizor (KO), an efficient stochastic optimization algorithm that adopts Kalman filter to make consistent estimation of the local gradient by solving an adaptive filtering problem. Our method reduces estimation variance in stochastic gradient descent by incorporating the historic state of the optimization. It aims to improve noisy gradient direction as well as accelerate the convergence of learning. We demonstrate the effectiveness of the proposed Kalman Optimizer under various optimization tasks where it is shown to achieve superior and robust performance. The code is available at https: //github.com/Adamdad/Filter-Gradient-Decent.
Year
DOI
Venue
2021
10.1109/ICASSP39728.2021.9414588
2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021)
Keywords
DocType
Citations 
Stochastic gradient descent, Kalman Filtering, Optimization
Conference
0
PageRank 
References 
Authors
0.34
2
1
Name
Order
Citations
PageRank
Xingyi Yang101.35