LEARNING-CURVES FOR ERROR MINIMUM AND MAXIMUM-LIKELIHOOD ALGORITHMS

被引:7
作者
KABASHIMA, Y
SHINOMOTO, S
机构
关键词
D O I
10.1162/neco.1992.4.5.712
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For the problem of dividing the space originally partitioned by a blurred boundary, every learning algorithm can make the probability of incorrect prediction of an individual example epsilon decrease with the number of training examples t. We address here the question of how the asymptotic form of epsilon(t) as well as its limit of convergence reflect the choice of learning algorithms. The error minimum algorithm is found to exhibit rather slow convergence of epsilon(t) to its lower bound epsilon0, epsilon(t) - epsilon0 approximately O(t-2/3). Even for the purpose of minimizing prediction error, the maximum likelihood algorithm can be utilized as an alternative. If the true probability distribution happens to be contained in the family of hypothetical functions, then the boundary estimated from the hypothetical distribution function eventually converges to the best choice. Convergence of the prediction error is then epsilon(t)-epsilon0 approximately O(t-1). If the true distribution is not available from the algorithm, however, the boundary generally does not converge to the best choice, but instead epsilon(t)-epsilon1 approximately +/- O(t-1/2), where epsilon1 > epsilon0 > 0.
引用
收藏
页码:712 / 719
页数:8
相关论文
共 6 条
[1]   4 TYPES OF LEARNING-CURVES [J].
AMARI, S ;
FUJITA, N ;
SHINOMOTO, S .
NEURAL COMPUTATION, 1992, 4 (04) :605-618
[2]   What Size Net Gives Valid Generalization? [J].
Baum, Eric B. ;
Haussler, David .
NEURAL COMPUTATION, 1989, 1 (01) :151-160
[3]  
HAUSSLER D, 1991, UCSCCRL9102
[4]   A STATISTICAL APPROACH TO LEARNING AND GENERALIZATION IN LAYERED NEURAL NETWORKS [J].
LEVIN, E ;
TISHBY, N ;
SOLLA, SA .
PROCEEDINGS OF THE IEEE, 1990, 78 (10) :1568-1574
[5]  
Rissanen J, 1989, STOCHASTIC COMPLEXIT
[6]   A THEORY OF THE LEARNABLE [J].
VALIANT, LG .
COMMUNICATIONS OF THE ACM, 1984, 27 (11) :1134-1142