CLASSIFICATION OF LINEARLY NONSEPARABLE PATTERNS BY LINEAR THRESHOLD ELEMENTS

被引:19
作者
ROYCHOWDHURY, VP
SIU, KY
KAILATH, T
机构
[1] UNIV CALIF IRVINE,DEPT ELECT & COMP ENGN,IRVINE,CA 92717
[2] STANFORD UNIV,INFORMAT SYST LAB,STANFORD,CA 94305
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1995年 / 6卷 / 02期
基金
美国国家科学基金会;
关键词
D O I
10.1109/72.363468
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning and convergence properties of linear threshold elements or perceptrons are well understood for the case where the input vectors (or the training sets) to the perceptron are linearly separable. Little is known, however, about the behavior of the perceptron learning algorithm when the training sets are linearly nonseparable. In this paper we present the first known results on the structure of linearly nonseparable training sets and on the behavior of perceptrons when the set of input vectors is linearly nonseparable. More precisely, we show that using the well known perceptron learning algorithm a linear threshold element can learn the input vectors that are provably learnable, and identify those vectors that cannot be learned without committing errors. We also show how a linear threshold element can be used to learn large linearly separable subsets of any given nonseparable training set. In order to develop our results, we first establish formal characterizations of linearly nonseparable training sets and define learnable structures for such patterns. We also prove computational complexity results for the related learning problems. Next, based on such characterizations, we show that a perceptron does the best one can expect for linearly nonseparable sets of input vectors and learns as much as is theoretically possible.
引用
收藏
页码:318 / 331
页数:14
相关论文
共 17 条
[1]  
ABUMOSTAFA YS, 1988, IEEE T INFORM THEORY, V32, P513
[2]   TRAINING A 3-NODE NEURAL NETWORK IS NP-COMPLETE [J].
BLUM, AL ;
RIVEST, RL .
NEURAL NETWORKS, 1992, 5 (01) :117-127
[3]   GEOMETRICAL AND STATISTICAL PROPERTIES OF SYSTEMS OF LINEAR INEQUALITIES WITH APPLICATIONS IN PATTERN RECOGNITION [J].
COVER, TM .
IEEE TRANSACTIONS ON ELECTRONIC COMPUTERS, 1965, EC14 (03) :326-&
[4]  
Garey M.R., 1979, COMPUTERS INTRACTABI, V174
[5]   ADAPTIVE HO-KASHYAP RULES FOR PERCEPTRON TRAINING [J].
HASSOUN, MH ;
JING, S .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (01) :51-61
[6]  
HASSOUN MH, 1986, THESIS WAYNE STATE U
[7]  
Judd J. S., 1990, P 1 ANN WORKSH COMP
[8]  
Kailath, 1995, DISCRETE NEURAL COMP
[9]  
McCulloch Warren S., 1943, BULL MATH BIOPHYS, V5, P115, DOI 10.1007/BF02478259
[10]  
MEGIDDO N, 1986, IBM RJ5252 ALM RES C