Automatic model selection for the optimization of SVM kernels

作者:

Highlights:

摘要

This approach aims to optimize the kernel parameters and to efficiently reduce the number of support vectors, so that the generalization error can be reduced drastically. The proposed methodology suggests the use of a new model selection criterion based on the estimation of the probability of error of the SVM classifier. For comparison, we considered two more model selection criteria: GACV (‘Generalized Approximate Cross-Validation’) and VC (‘Vapnik-Chernovenkis’) dimension. These criteria are algebraic estimates of upper bounds of the expected error. For the former, we also propose a new minimization scheme. The experiments conducted on a bi-class problem show that we can adequately choose the SVM hyper-parameters using the empirical error criterion. Moreover, it turns out that the criterion produces a less complex model with fewer support vectors. For multi-class data, the optimization strategy is adapted to the one-against-one data partitioning. The approach is then evaluated on images of handwritten digits from the USPS database.

论文关键词:Model selection,SVM,Kernel,Empirical error,VC,GACV

论文评审过程:Received 21 September 2004, Revised 23 March 2005, Accepted 23 March 2005, Available online 23 May 2005.

论文官网地址:https://doi.org/10.1016/j.patcog.2005.03.011