View a PDF of the paper titled Hidden Markov Neural Networks, by Lorenzo Rimella and Nick Whiteley
Abstract:We define an evolving in-time Bayesian neural network called a Hidden Markov Neural Network, which addresses the crucial challenge in time-series forecasting and continual learning: striking a balance between adapting to new data and appropriately forgetting outdated information. This is achieved by modelling the weights of a neural network as the hidden states of a Hidden Markov model, with the observed process defined by the available data. A filtering algorithm is employed to learn a variational approximation of the evolving-in-time posterior distribution over the weights. By leveraging a sequential variant of Bayes by Backprop, enriched with a stronger regularization technique called variational DropConnect, Hidden Markov Neural Networks achieve robust regularization and scalable inference. Experiments on MNIST, dynamic classification tasks, and next-frame forecasting in videos demonstrate that Hidden Markov Neural Networks provide strong predictive performance while enabling effective uncertainty quantification.
Submission history
From: Lorenzo Rimella [view email]
[v1]
Wed, 15 Apr 2020 09:18:18 UTC (1,003 KB)
[v2]
Wed, 24 Jun 2020 14:29:17 UTC (3,475 KB)
[v3]
Thu, 16 Jan 2025 08:32:50 UTC (7,485 KB)
Source link
lol