Maxi-min margin machine: Learning large margin classifiers locally and globally

被引:57
作者
Huang, Kaizhu [1 ]
Yang, Haiqin [2 ]
King, Irwin [3 ]
Lyu, Michael R.
机构
[1] Fujitsu Res & Dev Ctr Co Ltd, Informat Technol Lab, Beijing 100016, Peoples R China
[2] Titanium Technol Ltd, Shenzhen 518000, Peoples R China
[3] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Hong Kong, Hong Kong, Peoples R China
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 02期
关键词
classification; kernel methods; large margin; learning locally and globally; second-order cone programming;
D O I
10.1109/TNN.2007.905855
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
In this paper, we propose a novel large margin classifier, called the maxi-min margin machine (M-4). This model learns the decision boundary both locally and globally. In comparison, other large margin classifiers construct separating hyperplanes only either locally or globally. For example, a state-of-the-art large margin classifier, the support vector machine (SVM), considers data only locally, while another significant model, the minimax probability machine (MPM), focuses on building the decision hyperplane exclusively based on the global information. As a major contribution, we show that SVM yields the same solution as M-4 when data satisfy certain conditions, and MPM can be regarded as a relaxation model of M-4. Moreover, based on our proposed local and global view of data, another popular model, the linear discriminant analysis, can easily be interpreted and extended as well. We describe the M-4 model definition, provide a geometrical interpretation, present theoretical justifications, and propose a practical sequential conic programming method to solve the optimization problem. We also show how to exploit Mercer kernels to extend M-4 for nonlinear classifications. Furthermore, we perform a series of evaluations on both synthetic data sets and real-world benchmark data sets. Comparison with SVM and MPM demonstrates the advantages of our new model.
引用
收藏
页码:260 / 272
页数:13
相关论文
共 27 条
[1]
[Anonymous], 1999, The Nature Statist. Learn. Theory
[2]
[Anonymous], 1998, Encyclopedia of Biostatistics
[3]
Blake C.L., 1998, UCI repository of machine learning databases
[4]
BREIMAN L, 1997, 460 U CAL STAT DEPT
[5]
Fukunaga K., 1990, INTRO STAT PATTERN R
[6]
Huang KY, 2004, LECT NOTES COMPUT SC, V3181, P401
[7]
Huang KZ, 2004, J MACH LEARN RES, V5, P1253
[8]
Ivanov VK, 1962, SOV MATH DOKL, V145, P981
[9]
KOHAVI R, 1995, P 14 INT JOINT C ART, P338
[10]
Lanckriet G. R. G., 2003, Journal of Machine Learning Research, V3, P555, DOI 10.1162/153244303321897726