RATES OF CONVERGENCE IN ACTIVE LEARNING

被引:78
作者
Hanneke, Steve [1 ]
机构
[1] Carnegie Mellon Univ, Dept Stat, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
Active learning; sequential design; selective sampling; statistical learning theory; oracle inequalities; model selection; classification; EMPIRICAL PROCESSES; SAMPLE MODULI; INEQUALITIES; BOUNDS;
D O I
10.1214/10-AOS843
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We study the rates of convergence in generalization error achievable by active learning under various types of label noise. Additionally, we study the general problem of model selection for active learning with a nested hierarchy of hypothesis classes and propose an algorithm whose error rate provably converges to the best achievable error among classifiers in the hierarchy at a rate adaptive to both the complexity of the optimal classifier and the noise conditions. In particular, we state sufficient conditions for these rates to be dramatically faster than those achievable by passive learning.
引用
收藏
页码:333 / 361
页数:29
相关论文
共 33 条
[1]   SAMPLE MODULI FOR SET-INDEXED GAUSSIAN-PROCESSES [J].
ALEXANDER, KS .
ANNALS OF PROBABILITY, 1986, 14 (02) :598-611
[2]   RATES OF GROWTH AND SAMPLE MODULI FOR WEIGHTED EMPIRICAL PROCESSES INDEXED BY SETS [J].
ALEXANDER, KS .
PROBABILITY THEORY AND RELATED FIELDS, 1987, 75 (03) :379-423
[3]   PROBABILITY-INEQUALITIES FOR EMPIRICAL PROCESSES AND A LAW OF THE ITERATED LOGARITHM [J].
ALEXANDER, KS .
ANNALS OF PROBABILITY, 1984, 12 (04) :1041-1067
[4]  
ALEXANDER KS, 1985, P BERKELEY C HONOR J, V2, P475
[5]  
[Anonymous], 2007, Advances in Neural Information Processing Systems
[6]  
[Anonymous], P 17 INT C ALG LEARN
[7]  
Anthony Martin, 1999, Neural network learning: Theoretical foundations
[8]  
Balcan M., 2006, P 23 INT C MACH LEAR, P65
[9]  
Balcan M.-F., 2008, P 21 C LEARN THEOR
[10]   Margin based active learning [J].
Balcan, Maria-Florina ;
Broder, Andrei ;
Zhang, Tong .
LEARNING THEORY, PROCEEDINGS, 2007, 4539 :35-+