Scaling up support vector data description by using core-sets

被引:18
作者
Chu, CS [1 ]
Tsang, IW [1 ]
Kwok, JT [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Comp Sci, Hong Kong, Hong Kong, Peoples R China
来源
2004 IEEE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS | 2004年
关键词
D O I
10.1109/IJCNN.2004.1379943
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Support vector data description (SVDD) is a powerful kernel method that has been commonly used for novelty detection. While its quadratic programming formulation has the important computational advantage of avoiding the problem of local minimum, this has a runtime complexity of O(N-3), where N is the number of training patterns. It thus becomes prohibitive when the data set is large. Inspired from the use of core-sets in approximating the minimum enclosing ball problem in computational geometry, we propose in this paper an approximation method that allows SVDD to scale better to larger data sets. Most importantly, the proposed method has a running time that is only linear in N. Experimental results on two large real-world data sets demonstrate that the proposed method can handle data sets that are much larger than those that can be handled by standard SVDD packages, while its approximate solution still attains equally good, or sometimes even better, novelty detection performance.
引用
收藏
页码:425 / 430
页数:6
相关论文
共 21 条
  • [1] Badoiu M, 2003, SIAM PROC S, P801
  • [2] Badoiu M, 2002, P 34 ANN ACM S THEOR, P250, DOI DOI 10.1145/509907.509947
  • [3] The use of the area under the roc curve in the evaluation of machine learning algorithms
    Bradley, AP
    [J]. PATTERN RECOGNITION, 1997, 30 (07) : 1145 - 1159
  • [4] CAMPBELL C, 2002, ADV NEURAL INFORMATI, V14
  • [5] LIBSVM: A Library for Support Vector Machines
    Chang, Chih-Chung
    Lin, Chih-Jen
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2011, 2 (03)
  • [6] Cristianini N., 2000, An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods
  • [7] HAYTON P, 2001, ADV NEURAL INFORMATI, V13
  • [8] Kumar P, 2003, SIAM PROC S, P45
  • [9] LANCKRIET G, 2003, IMA WORKSH SEM PROGR
  • [10] LANCKRIET G, 2003, ADV NEURAL INFORMATI, V15