PERFECT LOSS OF GENERALIZATION DUE TO NOISE IN K=2 PARITY MACHINES

被引:14
作者
KABASHIMA, Y [1 ]
机构
[1] KYOTO UNIV,DEPT PHYS,KYOTO 606,JAPAN
来源
JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL | 1994年 / 27卷 / 06期
关键词
D O I
10.1088/0305-4470/27/6/017
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
Learning in a specific type of multilayer network referred to as a K = 2 parity machine is studied in the limit that both the system size N and the number of examples m become infinite while the ratio alpha = m/N remains finite, The machine consists of K = 2 hidden units with non-overlapping receptive fields each of size N/2. The output is the sign of the product of the two hidden units for each input. We investigate incremental learning by empirically using a least-action algorithm in the following two learning paradigms. In the first, it is assumed that each example is transmitted perfectly to a student. We show that an ability to generalize emerges as the rescaled length of the connection vector l reaches a critical value l(c). Further, we show that a student can identify the target exactly in the limit alpha --> infinity, where the prediction error epsilon decreases to zero as epsilon approximately 0.441alpha-1/3. In the second paradigm, we examine what happens if each teacher signal is reversed to the opposite sign at a noise rate lambda. For small lambda, it is found that the prediction error converges to a finite value of O (square-root lambda) in O (lambda-3/2) iterations. However, for a noise rate beyond a critical value lambda(c) approximately 0.175, the student cannot acquire any generalization ability even as a --> infinity.
引用
收藏
页码:1917 / 1927
页数:11
相关论文
共 12 条
[1]   4 TYPES OF LEARNING-CURVES [J].
AMARI, S ;
FUJITA, N ;
SHINOMOTO, S .
NEURAL COMPUTATION, 1992, 4 (04) :605-618
[2]   The Perceptron Algorithm is Fast for Nonmalicious Distributions [J].
Baum, Eric B. .
NEURAL COMPUTATION, 1990, 2 (02) :248-260
[3]   MEMORIZATION WITHOUT GENERALIZATION IN A MULTILAYERED NEURAL NETWORK [J].
HANSEL, D ;
MATO, G ;
MEUNIER, C .
EUROPHYSICS LETTERS, 1992, 20 (05) :471-476
[4]   LEARNING-CURVES FOR ERROR MINIMUM AND MAXIMUM-LIKELIHOOD ALGORITHMS [J].
KABASHIMA, Y ;
SHINOMOTO, S .
NEURAL COMPUTATION, 1992, 4 (05) :712-719
[5]  
KABASHIMA Y, 1993, 6TH P ACM C COMP LEA, P446
[6]   A NEW POLYNOMIAL-TIME ALGORITHM FOR LINEAR-PROGRAMMING [J].
KARMARKAR, N .
COMBINATORICA, 1984, 4 (04) :373-395
[7]  
KAWANABE M, 1993, ESTIMATION NEURONAL
[8]  
MINSKY M, 1989, PERCEPTRONS
[9]  
MITCHISON GJ, 1989, BIOL CYBERN, V60, P345, DOI 10.1007/BF00204772
[10]  
Nilsson N.J., 1965, LEARNING MACHINES