Tracking the best disjunction

被引:41
作者
Auer, P
Warmuth, MK
机构
[1] Graz Univ Technol, Inst Theoret Comp Sci, A-8010 Graz, Austria
[2] Univ Calif Santa Cruz, Dept Comp Sci, Santa Cruz, CA 95064 USA
关键词
on-line learning; prediction; concept drift; WINNOW; computational learning theory; amortized analysis;
D O I
10.1023/A:1007472513967
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
Littlestone developed a simple deterministic on-line learning algorithm for learning k-literal disjunctions. This algorithm (called WINNOW) keeps one weight for each of the n variables and does multiplicative updates to its weights. We develop a randomized version of WINNOW and prove bounds for an adaptation of the algorithm for the case when the disjunction may change over time. In this case a possible target disjunction schedule T is a sequence of disjunctions (one per trial) and the shift size is the total number of literals that are added/removed from the disjunctions as one progresses through the sequence. We develop an algorithm that predicts nearly as well as the best disjunction schedule for an arbitrary sequence of examples. This algorithm that allows us to track the predictions of the best disjunction is hardly more complex than the original version. However, the amortized analysis needed for obtaining worst-case mistake bounds requires new techniques. In some cases our lower bounds show that the upper bounds of our algorithm have the right constant in front of the leading term in the mistake bound and almost the right constant in front of the second leading term. Computer experiments support our theoretical Endings.
引用
收藏
页码:127 / 150
页数:24
相关论文
共 21 条
[1]
Auer P, 1995, AN S FDN CO, P312, DOI 10.1109/SFCS.1995.492487
[2]
Auer P., 1993, Proceeding of the Sixth Annual ACM Conference on Computational Learning Theory, P253, DOI 10.1145/168304.168345
[3]
On-line prediction and conversion strategies [J].
CesaBianchi, N ;
Freund, Y ;
Helmbold, DP ;
Warmuth, MK .
MACHINE LEARNING, 1996, 25 (01) :71-110
[4]
CESABIANCHI N, 1997, IN PRESS J ACM
[5]
Cover T. M., 1965, PROC 4 PRAGUE C INFO, P263
[6]
Duda R. O., 1973, PATTERN CLASSIFICATI, V3
[7]
HAUSSLER D, 1995, IN PRESS IEEE T INFO
[8]
Haykin S, 1998, NEURAL NETWORKS COMP
[9]
Predicting nearly as well as the best pruning of a decision tree [J].
Helmbold, DP ;
Schapire, RE .
MACHINE LEARNING, 1997, 27 (01) :51-68
[10]
HERBSTER M, 1998, IN PRESS MACHINE LEA