Title
Continual Learning with Bayesian Neural Networks for Non-Stationary Data
Abstract
This work addresses continual learning for non-stationary data, using Bayesian neural networks and memory-based online variational Bayes. We represent the posterior approximation of the network weights by a diagonal Gaussian distribution and a complementary memory of raw data. This raw data corresponds to likelihood terms that cannot be well approximated by the Gaussian. We introduce a novel method for sequentially updating both components of the posterior approximation. Furthermore, we propose Bayesian forgetting and a Gaussian diffusion process for adapting to non-stationary data. The experimental results show that our update method improves on existing approaches for streaming data. Additionally, the adaptation methods lead to better predictive performance for non-stationary data.
Year
Venue
Keywords
2020
ICLR
Continual Learning, Online Variational Bayes, Non-Stationary Data, Bayesian Neural Networks, Variational Inference, Lifelong Learning, Concept Drift, Episodic Memory
DocType
Citations 
PageRank 
Conference
0
0.34
References 
Authors
31
5
Name
Order
Citations
PageRank
Richard Kurle102.70
Botond Cseke219311.55
Alexej Klushyn301.69
Patrick van der Smagt418824.23
Stephan G眉nnemann583369.26