Specification and simulation of statistical query algorithms for efficiency and noise tolerance

被引:17
作者
Aslam, JA [1 ]
Decatur, SE
机构
[1] Dartmouth Coll, Dept Comp Sci, Hanover, NH 03755 USA
[2] MIT, Comp Sci Lab, Cambridge, MA 02139 USA
基金
美国国家科学基金会;
关键词
D O I
10.1006/jcss.1997.1558
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
A recent innovation in computational learning theory is the statistical query (SQ) model. The advantage of specifying learning algorithms in this model is that SO algorithms can be simulated in the probably approximately correct (PAC) model, both in the absence and in the presence of noise. However, simulations of SO algorithms in the PAC model have non-optimal time and sample complexities. In this paper, we introduce a new method for specifying statistical query algorithms based on a type of relative error and provide simulations in the noise-free and noise-tolerant PAC models which yield more efficient algorithms. Requests for estimates of statistics in this new model take the following form: "Return an estimate of the statistic within a 1 +/- mu factor, or return perpendicular to, promising that the statistic is less than theta." In addition to showing that this is a very natural language for specifying learning algorithms, we also shaw that this new specification is polynomially equivalent to standard SQ, and thus, known learnability and hardness results for statistical query learning are preserved. We then give highly efficient PAC simulations of relative error SQ algorithms. We show that the learning algorithms obtained by simulating efficient relative error SQ algorithms both in the absence of noise and in the presence of malicious noise have roughly optimal sample complexity. We also show that the simulation of efficient relative error SQ algorithms in the presence of classification noise yields learning algorithms at least as efficient as those obtained through standard methods, and in some cases improved, roughly optimal results are achieved. The sample complexities for all of these simulations are based on the d(v) metric, which is a type of relative error metric useful for quantities which are small or even zero. We show that uniform convergence with respect to the d(v) metric yields "uniform convergence" with respect to (mu, theta) accuracy. Finally, while we show that many specific learning algorithms can be written as highly efficient relative error SO algorithms, we also show, in fact, that all SO algorithms can be written efficiently by proving general upper bounds on the complexity of (mu, theta) queries as a function of the accuracy parameter epsilon. As a consequence of this result, we give general upper bounds on the complexity of learning algorithms achieved through the use of relative error SO algorithms and the simulations described above. (C) 1998 Academic Press.
引用
收藏
页码:191 / 208
页数:18
相关论文
共 20 条
[1]   FAST PROBABILISTIC ALGORITHMS FOR HAMILTONIAN CIRCUITS AND MATCHINGS [J].
ANGLUIN, D ;
VALIANT, LG .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1979, 18 (02) :155-193
[2]  
Aslam J. A., 1993, Proceedings. 34th Annual Symposium on Foundations of Computer Science (Cat. No.93CH3368-8), P282, DOI 10.1109/SFCS.1993.366859
[3]   General bounds on statistical query learning and PAC learning with noise via hypothesis boosting [J].
Aslam, JA ;
Decatur, SE .
INFORMATION AND COMPUTATION, 1998, 141 (02) :85-118
[4]   On the sample complexity of noise-tolerant learning [J].
Aslam, JA ;
Decatur, SE .
INFORMATION PROCESSING LETTERS, 1996, 57 (04) :189-195
[5]  
Blum A, 1994, P 26 ANN ACM S THEOR
[6]  
BLUMER A, 1989, J ASSOC COMPUT MACH, V36, P829
[7]  
DECATUR S, 1995, P 8 ANN ACM WORKSH C
[8]  
DECATUR S, 1995, P 5 INT WORKSH ART I, P175
[9]  
Decatur S. E., 1993, Proceeding of the Sixth Annual ACM Conference on Computational Learning Theory, P262, DOI 10.1145/168304.168346
[10]   A GENERAL LOWER BOUND ON THE NUMBER OF EXAMPLES NEEDED FOR LEARNING [J].
EHRENFEUCHT, A ;
HAUSSLER, D ;
KEARNS, M ;
VALIANT, L .
INFORMATION AND COMPUTATION, 1989, 82 (03) :247-261