UNSUPERVISED MUTUAL INFORMATION CRITERION FOR ELIMINATION OF OVERTRAINING IN SUPERVISED MULTILAYER NETWORKS

被引:54
作者
DECO, G
FINNOFF, W
ZIMMERMANN, HG
机构
关键词
D O I
10.1162/neco.1995.7.1.86
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Controlling the network complexity in order to prevent overfitting is one of the major problems encountered when using neural network models to extract the structure from small data sets. In this paper we present a network architecture designed for use with a cost function that includes a novel complexity penalty term. In this architecture the outputs of the hidden units are strictly positive and sum to one, and their outputs are defined as the probability that the actual input belongs to a certain class formed during learning. The penalty term expresses the mutual information between the inputs and the extracted classes. This measure effectively describes the network complexity with respect to the given data in an unsupervised fashion. The efficiency of this architechture/penalty-term when combined with back-propagation training, is demonstrated on a real world economic time series forecasting problem. The model was also applied to the bench-mark sunspot data and to a synthetic data set from the statistics community.
引用
收藏
页码:86 / 107
页数:22
相关论文
共 29 条