Title
Differentiable PAC-Bayes Objectives with Partially Aggregated Neural Networks
Abstract
We make two related contributions motivated by the challenge of training stochastic neural networks, particularly in a PAC-Bayesian setting: (1) we show how averaging over an ensemble of stochastic neural networks enables a new class of partially-aggregated estimators, proving that these lead to unbiased lower-variance output and gradient estimators; (2) we reformulate a PAC-Bayesian bound for signed-output networks to derive in combination with the above a directly optimisable, differentiable objective and a generalisation guarantee, without using a surrogate loss or loosening the bound. We show empirically that this leads to competitive generalisation guarantees and compares favourably to other methods for training such networks. Finally, we note that the above leads to a simpler PAC-Bayesian training scheme for sign-activation networks than previous work.
Year
DOI
Venue
2021
10.3390/e23101280
ENTROPY
Keywords
DocType
Volume
statistical learning theory, PAC-Bayes theory, deep learning
Journal
23
Issue
ISSN
Citations 
10
1099-4300
0
PageRank 
References 
Authors
0.34
0
2
Name
Order
Citations
PageRank
Felix Biggs100.68
Benjamin Guedj298.82