Tracking the best expert

被引:270
作者
Herbster, M [1 ]
Warmuth, MK [1 ]
机构
[1] Univ Calif Santa Cruz, Dept Comp Sci, Santa Cruz, CA 95064 USA
关键词
on-line learning; amortized analysis; multiplicative updates; shifting; experts;
D O I
10.1023/A:1007424614876
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
We generalize the recent relative loss bounds for on-line algorithms where the additional loss of the algorithm on the whole sequence of examples over the loss of the best expert is bounded. The generalization allows the sequence to be partitioned into segments. and the goal is to bound the additional loss of the algorithm over the sum of the losses of the best experts for each segment. This is to model situations in which the examples change and different experts are best for certain segments of the sequence of examples. In the single segment case, the additional loss is proportional to log n, where n is the number of experts and the constant of proportionality depends on the loss function. Our algorithms do not produce the best partition: however the loss bound shows that our predictions are close to those of the best partition. When the number of segments is k + 1 and the sequence is of length l. we can bound the additional loss of our algorithm over the best partition by O(k log n + k log(l/k)). For the case when the loss per trial is bounded by one, we obtain an algorithm whose additional loss over the loss of the best partition is independent of the length of the sequence. The additional loss becomes O(k log n + k log(L/k)), where L is the loss of the best partition with k + 1 segments. Our algorithms for tracking the predictions of the best expert are simple adaptations of Vovk's original algorithm for the single best expert case. As in the original algorithms, we keep one weight per expert, and spend O(1) lime per weight in each trial.
引用
收藏
页码:151 / 178
页数:28
相关论文
共 19 条
[1]
BLUM A, 1997, P 10 ANN WORKSH COMP
[2]
How to use expert advice [J].
CesaBianchi, N ;
Freund, Y ;
Haussler, D ;
Helmbold, DP ;
Schapire, RE ;
Warmuth, MK .
JOURNAL OF THE ACM, 1997, 44 (03) :427-485
[3]
Cover T. M., 2005, ELEM INF THEORY, DOI 10.1002/047174882X
[4]
UNIVERSAL PREDICTION OF INDIVIDUAL SEQUENCES [J].
FEDER, M ;
MERHAV, N ;
GUTMAN, M .
IEEE TRANSACTIONS ON INFORMATION THEORY, 1992, 38 (04) :1258-1270
[5]
A decision-theoretic generalization of on-line learning and an application to boosting [J].
Freund, Y ;
Schapire, RE .
JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 1997, 55 (01) :119-139
[6]
Freund Yoav, 1997, P 29 ANN ACM S THEOR
[7]
HAUSSLER D, 1998, IN PRESS IEEE T INFO
[8]
HELMBOLD DP, 1996, P 2 ANN ACM INT C MO
[9]
HELMBOLD DP, 1995, P 1995 NEUR INF PROC, P309
[10]
HERBSTER M, 1997, UNPUB