Extraction of comprehensive symbolic rules from a multi-layer perceptron

被引:5
作者
Avner, S
机构
[1] LAFORIA-CNRS URA 1095 IBP, Université Paris VI, 75252 Paris Cedex 05
关键词
learning from examples; neural-network interpretation; generalization; extraction of comprehensible symbolic rules;
D O I
10.1016/0952-1976(96)00004-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper introduces a system that extracts comprehensible symbolic rules from a multi-layer perceptron. Once the network has been trained in the usual manner, the training set is presented again, and the actual activations of the units recorded. Logical rules, corresponding to the logical combinations of the incoming signals, are extracted at each activated unit. This procedure is used for all examples belonging to the training set. Thus a set of rules is obtained which accounts for all logical steps taken by the network to process all known input patterns. Furthermore, it is shown that if some symbolic meaning were associated with every input unit, then the hidden units, which have formed concepts in order to deal with recurrent features in the input data, possess some symbolic meaning too! The algorithm described in this paper allows the recognition or the understandability of these concepts: they are found to be reducible to conjunctions and negations of the human input concepts. The rules can also be recombined in different ways, thus constituting some limited bur sound generalization of the training set. Neural networks could learn concepts about domains where little theory was known but where many examples were available. Yet, because their knowledge was stored in the synaptic strengths under numerical form, it was difficult to comprehend what they had discovered. This system therefore provides some means of accessing the information contained inside the network. Copyright (C) 1996 Elsevier Science Ltd
引用
收藏
页码:137 / 143
页数:7
相关论文
共 18 条
[1]  
Barlow H B, 1972, Perception, V1, P371, DOI 10.1068/p010371
[2]  
BEALE R., 1990, Neural Computing: An Introduction, DOI DOI 10.1887/0852742622
[3]  
CHAN LW, 1991, PROCEEDINGS : NINTH NATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOLS 1 AND 2, P578
[4]  
Denker J., 1987, Complex Systems, V1, P877
[5]  
FU LM, 1991, 9TH P NAT C ART INT, P590
[6]   CONNECTIONIST EXPERT SYSTEMS [J].
GALLANT, SI .
COMMUNICATIONS OF THE ACM, 1988, 31 (02) :152-169
[7]   WHAT CONNECTIONIST MODELS LEARN - LEARNING AND REPRESENTATION IN CONNECTIONIST NETWORKS [J].
HANSON, SJ ;
BURR, DJ .
BEHAVIORAL AND BRAIN SCIENCES, 1990, 13 (03) :471-488
[8]  
Hertz J., 1991, Introduction to the Theory of Neural Computation
[9]  
Langley P., 1989, Machine Learning, V3, P253
[10]  
LeCun Y., 1990, Advances in neural information processing systems, P598