Title
Neural network training using ℓ1-regularization and bi-fidelity data
Abstract
With the capability of accurately representing a functional relationship between the inputs of a physical system's model and output quantities of interest, neural networks have become popular for surrogate modeling in scientific applications. However, as these networks are over-parameterized, their training often requires a large amount of data. To prevent overfitting and improve generalization error, regularization based on, e.g., ℓ1- and ℓ2-norms of the parameters is applied. Similarly, multiple connections of the network may be pruned to increase sparsity in the network parameters. In this paper, we explore the effects of sparsity promoting ℓ1-regularization on training neural networks when only a small training dataset from a high-fidelity model is available. As opposed to standard ℓ1-regularization that is known to be inadequate, we consider two variants of ℓ1-regularization informed by the parameters of an identical network trained using data from lower-fidelity models of the problem at hand. These bi-fidelity strategies are generalizations of transfer learning of neural networks that uses the parameters learned from a large low-fidelity dataset to efficiently train networks for a small high-fidelity dataset. We also compare the bi-fidelity strategies with two ℓ1-regularization methods that only use the high-fidelity dataset but with similar computational cost for generating the training datasets. Three numerical examples for propagating uncertainty through physical systems are used to show that the proposed bi-fidelity ℓ1-regularization strategies produce errors that are one order of magnitude smaller than those of networks trained only using datasets from the high-fidelity models.
Year
DOI
Venue
2022
10.1016/j.jcp.2022.111010
Journal of Computational Physics
Keywords
DocType
Volume
Neural networks,Scientific machine learning,Transfer learning,Weighted ℓ1-regularization,Uncertainty propagation
Journal
458
ISSN
Citations 
PageRank 
0021-9991
0
0.34
References 
Authors
0
2
Name
Order
Citations
PageRank
Subhayan De100.34
Alireza Doostan218815.57