Title
Temporal Dynamics of Generalization in Neural Networks
Abstract
This paper presents a rigorous characterization of how a general nonlinear learning machine generalizes during the training process when it is trained on a random sample using a gradient descent algorithm based on reduction of training error. It is shown, in particular, that best generalization performance occurs, in general, before the global minimum of the training error is achieved. The different roles played by the complexity of the machine class and the complexity of the specific machine in the class during learning are also precisely demarcated.
Year
Venue
Keywords
1994
NIPS
neural network
DocType
Citations 
PageRank 
Conference
6
0.69
References 
Authors
1
2
Name
Order
Citations
PageRank
Changfeng Wang160.69
Santosh S. Venkatesh238171.80