Paper ID | MLSP-41.3 |
Paper Title |
Kalman Optimizer for Consistent Gradient Descent |
Authors |
Xingyi Yang, University of California, San Diego, United States |
Session | MLSP-41: Deep Learning Optimization |
Location | Gather.Town |
Session Time: | Friday, 11 June, 11:30 - 12:15 |
Presentation Time: | Friday, 11 June, 11:30 - 12:15 |
Presentation |
Poster
|
Topic |
Machine Learning for Signal Processing: [MLR-DEEP] Deep learning techniques |
IEEE Xplore Open Preview |
Click here to view in IEEE Xplore |
Virtual Presentation |
Click here to watch in the Virtual Conference |
Abstract |
Deep neural networks (DNN) are typically optimized using stochastic gradient descent (SGD). However, the estimation of the gradient using stochastic samples tends to be noisy and unreliable, resulting in large gradient variance and bad convergence. In this paper, we propose Kalman Optimizor (KO), an efficient stochastic optimization algorithm that adopts Kalman filter to make consistent estimation of the local gradient by solving an adaptive filtering problem. Our method reduces estimation variance in stochastic gradient descent by incorporating the historic state of the optimization. It aims to improve noisy gradient direction as well as accelerate the convergence of learning. We demonstrate the effectiveness of the proposed Kalman Optimizer under various optimization tasks where it is shown to achieve superior and robust performance. The code is available at https://github.com/Adamdad/Filter-Gradient-Decent. |