Title
Model complexity control for regression using VC generalization bounds.
Abstract
It is well known that for a given sample size there exists a model of optimal complexity corresponding to the smallest prediction (generalization) error. Hence, any method for learning from finite samples needs to have some provisions for complexity control. Existing implementations of complexity control include penalization (or regularization), weight decay (in neural networks), and various greedy procedures (aka constructive, growing, or pruning methods). There are numerous proposals for determining optimal model complexity (aka model selection) based on various (asymptotic) analytic estimates of the prediction risk and on resampling approaches. Nonasymptotic bounds on the prediction risk based on Vapnik-Chervonenkis (VC)-theory have been proposed by Vapnik. This paper describes application of VC-bounds to regression problems with the usual squared loss. An empirical study is performed for settings where the VC-bounds can be rigorously applied, i.e., linear models and penalized linear models where the VC-dimension can be accurately estimated, and the empirical risk can be reliably minimized. Empirical comparisons between model selection using VC-bounds and classical methods are performed for various noise levels, sample size, target functions and types of approximating functions. Our results demonstrate the advantages of VC-based complexity control with finite samples.
Year
DOI
Venue
1999
10.1109/72.788648
IEEE Transactions on Neural Networks
Keywords
Field
DocType
finite sample,linear model,optimal model complexity,model selection,vc generalization bound,prediction risk,vc-based complexity control,aka model selection,sample size,model complexity control,complexity control,optimal complexity,generalization error,indexing terms,empirical risk minimization,machine learning,vc dimension,neural network,regression,predictive models,control systems,empirical study,minimisation,statistical analysis,neural networks,training data,risk analysis
Generalization,Linear model,Computer science,Model selection,Minimisation (psychology),Regularization (mathematics),Artificial intelligence,Artificial neural network,Resampling,Sample size determination,Machine learning
Journal
Volume
Issue
ISSN
10
5
1045-9227
Citations 
PageRank 
References 
65
10.12
3
Authors
4
Name
Order
Citations
PageRank
Vladimir Cherkassky17817891.92
X Shao219823.20
F. Mulier324542.81
Vladimir Vapnik4160753397.91