Title
In-Sample and Out-of-Sample Model Selection and Error Estimation for Support Vector Machines
Abstract
In-sample approaches to model selection and error estimation of support vector machines (SVMs) are not as widespread as out-of-sample methods, where part of the data is removed from the training set for validation and testing purposes, mainly because their practical application is not straightforward and the latter provide, in many cases, satisfactory results. In this paper, we survey some recent and not-so-recent results of the data-dependent structural risk minimization framework and propose a proper reformulation of the SVM learning algorithm, so that the in-sample approach can be effectively applied. The experiments, performed both on simulated and real-world datasets, show that our in-sample approach can be favorably compared to out-of-sample methods, especially in cases where the latter ones provide questionable results. In particular, when the number of samples is small compared to their dimensionality, like in classification of microarray data, our proposal can outperform conventional out-of-sample approaches such as the cross validation, the leave-one-out, or the Bootstrap methods.
Year
DOI
Venue
2012
10.1109/TNNLS.2012.2202401
IEEE Trans. Neural Netw. Learning Syst.
Keywords
Field
DocType
support vector machines,statistical analysis,learning artificial intelligence
Data mining,Computer science,Artificial intelligence,Structural risk minimization,Bootstrapping (electronics),Least squares support vector machine,Pattern recognition,Support vector machine,Model selection,Curse of dimensionality,Supervised learning,Cross-validation,Machine learning
Journal
Volume
Issue
ISSN
23
9
2162-237X
Citations 
PageRank 
References 
50
1.37
33
Authors
4
Name
Order
Citations
PageRank
Davide Anguita1100170.58
Alessandro Ghio266735.71
Luca Oneto383063.22
Sandro Ridella4677140.62