A new algorithm to design compact two-hidden-layer artificial neural networks

被引:76
作者
Islam, MM [1 ]
Murase, K [1 ]
机构
[1] Univ Fukui, Dept Human & Artificial Intelligence Syst, Fukui 9108507, Japan
关键词
constructive algorithm; pruning algorithm; weight freezing; generalization ability; training time;
D O I
10.1016/S0893-6080(01)00075-2
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper describes the cascade neural network design algorithm (CNNDA), a new algorithm for designing compact, two-hidden-layer artificial neural networks (ANNs). This algorithm determines an ANN's architecture with connection weights automatically. The design strategy used in the CNNDA was intended to optimize both the generalization ability and the training time of ANNs. In order to improve the generalization ability, the CNDDA uses a combination of constructive and pruning algorithms and bounded fan-ins of the hidden nodes. A new training approach, by which the input weights of a hidden node are temporarily frozen when its output does not change much after a few successive training cycles, was used in the CNNDA for reducing the computational cost and the training time. The CNNDA was tested on several benchmarks including the cancer, diabetes and character-recognition problems in ANNs. The experimental results show that the CNNDA can produce compact ANNs with good generalization ability and short training time in comparison with other algorithms. (C) 2001 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:1265 / 1278
页数:14
相关论文
共 30 条
[1]   EFFICIENT CLASSIFICATION FOR MULTICLASS PROBLEMS USING MODULAR NEURAL NETWORKS [J].
ANAND, R ;
MEHROTRA, K ;
MOHAN, CK ;
RANKA, S .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1995, 6 (01) :117-124
[2]  
[Anonymous], INT WORKSH COMB GEN
[3]  
Ash T., 1989, Connection Science, V1, P365, DOI 10.1080/09540098908915647
[4]  
Baluja S., 1994, REDUCING NETWORK DEP
[5]   DYNAMIC NODE ARCHITECTURE LEARNING - AN INFORMATION-THEORETIC APPROACH [J].
BARTLETT, EB .
NEURAL NETWORKS, 1994, 7 (01) :129-140
[6]   1ST-ORDER AND 2ND-ORDER METHODS FOR LEARNING - BETWEEN STEEPEST DESCENT AND NEWTON METHOD [J].
BATTITI, R .
NEURAL COMPUTATION, 1992, 4 (02) :141-166
[7]  
Bourlard H. A., 1994, Connectionist speech recognition: a hybrid approach
[8]  
Fahlman S. E., 1990, ADV NEURAL INFORMATI, P524, DOI DOI 10.1190/1.1821929
[9]   IMPROVING MODEL SELECTION BY NONCONVERGENT METHODS [J].
FINNOFF, W ;
HERGERT, F ;
ZIMMERMANN, HG .
NEURAL NETWORKS, 1993, 6 (06) :771-783
[10]   NEURAL NETWORKS AND THE BIAS VARIANCE DILEMMA [J].
GEMAN, S ;
BIENENSTOCK, E ;
DOURSAT, R .
NEURAL COMPUTATION, 1992, 4 (01) :1-58