Title
Evaluation of Performance Measures for SVR Hyperparameter Selection
Abstract
To obtain accurate modeling results, it is of primal importance to find optimal values for the hyperparameters in the Support Vector Regression (SVR) model. In general, we search for those parameters that minimize an estimate of the generalization error. In this study, we empirically investigate different performance measures found in the literature: k-fold cross-validation, the computationally intensive, but almost unbiased leave-one-out error, its upper bounds -radius/margin and span bound -, Vapnik's measure, which uses an estimate of the VC dimension, and the regularized risk functional itself. For each of the estimates we focus on accuracy, complexity and the presence of local minima. The latter significantly influences the applicability of gradient-based search techniques to determine the optimal parameters.
Year
DOI
Venue
2007
10.1109/IJCNN.2007.4371031
Orlando, FL
Keywords
Field
DocType
estimation theory,generalisation (artificial intelligence),gradient methods,regression analysis,search problems,support vector machines,Vapnik measure,generalization error estimation,gradient-based search technique,k-fold cross-validation,leave-one-out error,performance measure,support vector regression hyperparameter selection
VC dimension,Regression analysis,Artificial intelligence,Estimation theory,Mathematical optimization,Pattern recognition,Hyperparameter,Support vector machine,Maxima and minima,Generalization error,Mathematics,Leave-one-out error,Machine learning
Conference
ISSN
ISBN
Citations 
1098-7576 E-ISBN : 978-1-4244-1380-5
978-1-4244-1380-5
5
PageRank 
References 
Authors
0.49
10
3
Name
Order
Citations
PageRank
Koen Smets150.49
Brigitte M. Verdonk28727.05
Elsa M. Jordaan3213.27