Optimizing resources in model selection for support vector machine

被引:30
作者
Adankon, Mathias M. [1 ]
Cheriet, Mohamed [1 ]
机构
[1] ETS, Lab Imagery Vis & Artif Intelligence, Montreal, PQ H3C 1K3, Canada
关键词
model selection; SVM; kernel; hyperparameters; optimizing time;
D O I
10.1016/j.patcog.2006.06.012
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Tuning support vector machine (SVM) hyperparameters is an important step in achieving a high-performance learning machine. It is usually done by minimizing an estimate of generalization error based on the bounds of the leave-one-out (LOO) such as radius-margin bound and on the performance measures such as generalized approximate cross-validation (GACV), empirical error, etc. These usual automatic methods used to tune the hyperparameters require an inversion of the Gram-Schmidt matrix or a resolution of an extra-quadratic programming problem. In the case of a large data set these methods require the addition of huge amounts of memory and a long CPU time to the already significant resources used in SVM training. In this paper, we propose a fast method based on an approximation of the gradient of the empirical error, along with incremental learning, which reduces the resources required both in terms of processing time and of storage space. We tested our method on several benchmarks, which produced promising results confirming our approach. Furthermore, it is worth noting that the gain time increases when the data set is large. (c) 2006 Pattern Recognition Society. Published by Elsevier Ltd. All rights reserved.
引用
收藏
页码:953 / 963
页数:11
相关论文
共 28 条
[1]  
[Anonymous], 2004, KERNEL METHODS PATTE
[2]  
[Anonymous], 1982, ESTIMATION DEPENDENC
[3]   Automatic model selection for the optimization of SVM kernels [J].
Ayat, NE ;
Cheriet, M ;
Suen, CY .
PATTERN RECOGNITION, 2005, 38 (10) :1733-1745
[4]   Empirical error based optimization of SVM kernels: Application to digit image recognition. [J].
Ayat, NE ;
Cheriet, M ;
Suen, CY .
EIGHTH INTERNATIONAL WORKSHOP ON FRONTIERS IN HANDWRITING RECOGNITION: PROCEEDINGS, 2002, :292-297
[5]  
AYAT NE, 2003, THESIS U QUEBEC
[6]   Gradient-based optimization of hyperparameters [J].
Bengio, Y .
NEURAL COMPUTATION, 2000, 12 (08) :1889-1900
[7]  
Bishop CM., 1995, Neural networks for pattern recognition
[8]  
Chapelle O, 2000, ADV NEUR IN, V12, P230
[9]   Choosing multiple parameters for support vector machines [J].
Chapelle, O ;
Vapnik, V ;
Bousquet, O ;
Mukherjee, S .
MACHINE LEARNING, 2002, 46 (1-3) :131-159
[10]   Radius margin bounds for support vector machines with the RBF kernel [J].
Chung, KM ;
Kao, WC ;
Sun, CL ;
Wang, LL ;
Lin, CJ .
NEURAL COMPUTATION, 2003, 15 (11) :2643-2681