SemiBoost: Boosting for Semi-Supervised Learning

被引:250
作者
Mallapragada, Pavan Kumar [1 ]
Jin, Rong [1 ]
Jain, Anil K. [1 ]
Liu, Yi [1 ]
机构
[1] Michigan State Univ, Dept Comp Sci & Engn, E Lansing, MI 48823 USA
基金
美国国家科学基金会;
关键词
Machine learning; semi-supervised learning; semi-supervised improvement; manifold assumption; cluster assumption; boosting;
D O I
10.1109/TPAMI.2008.235
中图分类号
TP18 [人工智能理论];
学科分类号
140502 [人工智能];
摘要
Semi-supervised learning has attracted a significant amount of attention in pattern recognition and machine learning. Most previous studies have focused on designing special algorithms to effectively exploit the unlabeled data in conjunction with labeled data. Our goal is to improve the classification accuracy of any given supervised learning algorithm by using the available unlabeled examples. We call this as the Semi-supervised improvement problem, to distinguish the proposed approach from the existing approaches. We design a metasemi-supervised learning algorithm that wraps around the underlying supervised algorithm and improves its performance using unlabeled data. This problem is particularly important when we need to train a supervised learning algorithm with a limited number of labeled examples and a multitude of unlabeled examples. We present a boosting framework for semi-supervised learning, termed as SemiBoost. The key advantages of the proposed semi-supervised learning approach are: 1) performance improvement of any supervised learning algorithm with a multitude of unlabeled data, 2) efficient computation by the iterative boosting algorithm, and 3) exploiting both manifold and cluster assumption in training classification models. An empirical study on 16 different data sets and text categorization demonstrates that the proposed framework improves the performance of several commonly used supervised learning algorithms, given a large number of unlabeled examples. We also show that the performance of the proposed algorithm, SemiBoost, is comparable to the state-of-the-art semi-supervised learning algorithms.
引用
收藏
页码:2000 / 2014
页数:15
相关论文
共 34 条
[1]
[Anonymous], 2006, Semi-supervised learning
[2]
[Anonymous], Data Mining Practical Machine Learning Tools and Techniques with Java
[3]
[Anonymous], 2006, IEEE T NEURAL NETWOR
[4]
[Anonymous], 2001, P NEUR INF PROC SYST
[5]
[Anonymous], P 22 INT C MACH LEAR
[6]
BELKIN M, 2004, T200406 U CHIC DEP C
[7]
Bennett K.P., 2002, P 8 ACM SIGKDD INT C, P289
[8]
Bennett KP, 1999, ADV NEUR IN, V11, P368
[9]
Blum A., 1998, Proceedings of the Eleventh Annual Conference on Computational Learning Theory, P92, DOI 10.1145/279943.279962
[10]
Blum A, 2001, P 18 INT C MACH LEAR, P19, DOI DOI 10.1184/R1/6606860.V1