Input space versus feature space in kernel-based methods

被引:816
作者
Schölkopf, B
Mika, S
Burges, CJC
Knirsch, P
Müller, KR
Rätsch, G
Smola, AJ
机构
[1] GMD FIRST, D-12489 Berlin, Germany
[2] AT&T Bell Labs, Holmdel, NJ 07733 USA
[3] Max Planck Inst Biol Cybernet, D-72076 Tubingen, Germany
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1999年 / 10卷 / 05期
基金
澳大利亚研究理事会;
关键词
denoising; kernel methods; PCA; reduced set method; sparse representation; support vector machines;
D O I
10.1109/72.788641
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper collects some ideas targeted at advancing our understanding of the feature spaces associated with support vector (SV) kernel functions. We first discuss the geometry of feature space. in particular, we review what is known about the shape of the image of input space under the feature space map, and how this influences the capacity of SV methods. Following this, we describe how the metric governing the intrinsic geometry of the mapped surface can be computed in terms of the kernel, using the example of the class of inhomogeneous polynomial kernels, which are often used in SV pattern recognition. We then discuss the connection between feature space and input space by dealing with the question of how one can, given some vector in feature space, find a preimage (exact or approximate) in input space. We describe algorithms to tackle this issue, and show their utility in two applications of kernel methods.. First, we use it to reduce the computational complexity of SV decision functions; second, we combine it with the Kernel PCA algorithm, thereby constructing a nonlinear statistical denoising technique which is shown to perform well on real-world data.
引用
收藏
页码:1000 / 1017
页数:18
相关论文
共 38 条
[1]  
AIZERMAN MA, 1965, AUTOMAT REM CONTR+, V25, P821
[2]  
[Anonymous], [No title captured], DOI DOI 10.1023/A:1009715923555
[3]  
[Anonymous], THESIS TU BERLIN
[4]  
[Anonymous], P 13 INT C MACH LEAR
[5]  
[Anonymous], 1999, ADV NEURAL INFORM PR
[6]  
[Anonymous], 1995, HDB BRAIN THEORY NEU
[7]   THEORY OF REPRODUCING KERNELS [J].
ARONSZAJN, N .
TRANSACTIONS OF THE AMERICAN MATHEMATICAL SOCIETY, 1950, 68 (MAY) :337-404
[8]   Decision region approximation by polynomials or neural networks [J].
Blackmore, KL ;
Williamson, RC ;
Mareels, IMY .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1997, 43 (03) :903-907
[9]  
Boser B. E., 1992, Proceedings of the Fifth Annual ACM Workshop on Computational Learning Theory, P144, DOI 10.1145/130385.130401
[10]  
Burges CJC, 1999, ADVANCES IN KERNEL METHODS, P89