Training algorithm with incomplete data for feed-forward neural networks

被引:46
作者
Yoon, SY [1 ]
Lee, SY [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Elect Engn, Computat & Neural Syst Lab, Taejon 305701, South Korea
关键词
missing data; incomplete data; estimation; supervised learning;
D O I
10.1023/A:1018772122605
中图分类号
TP18 [人工智能理论];
学科分类号
081104 [模式识别与智能系统]; 0812 [计算机科学与技术]; 0835 [软件工程]; 1405 [智能科学与技术];
摘要
A new algorithm is developed to train feed-forward neural networks for non-linear input-to-output mappings with small incomplete data in arbitrary distributions. The developed Training-EStimation-Training (TEST) algorithm consists of 3 steps, i.e., (1) training with the complete portion of the training data set, (2) estimation of the missing attributes with the trained neural networks, and (3) re-training the neural networks with the whole data set. Error back propagation is still applicable to estimate the missing attributes. Unlike other training methods with missing data, it does not assume data distribution models which may not be appropriate for small training data. The developed TEST algorithm is first tested for the Iris benchmark data. By randomly removing some attributes from the complete data set and estimating the values latter, accuracy of the TEST algorithm is demonstrated. Then it is applied to the Diabetes benchmark data, of which about 50% contains missing attributes. Compared with other existing algorithms, the proposed TEST algorithm results in much better recognition accuracy for test data.
引用
收藏
页码:171 / 179
页数:9
相关论文
共 15 条
[1]
Information geometry of the EM and em algorithms for neural networks [J].
Amari, SI .
NEURAL NETWORKS, 1995, 8 (09) :1379-1408
[2]
BEALE EML, 1975, J ROY STAT SOC B MET, V37, P129
[3]
CHEN CF, 1986, BAYESIAN INFERENCE D, P355
[4]
MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[5]
Ghahramani Z., 1994, Advances in Neural Information Processing Systems, V6, P120
[6]
HEITJAN DF, 1991, APPL STAT-J ROY ST C, V40, P13
[7]
USING ADDITIVE NOISE IN BACK-PROPAGATION TRAINING [J].
HOLMSTROM, L ;
KOISTINEN, P .
IEEE TRANSACTIONS ON NEURAL NETWORKS, 1992, 3 (01) :24-38
[8]
Merging back-propagation and Hebbian learning rules for robust classifications [J].
Jeong, DG ;
Lee, SY .
NEURAL NETWORKS, 1996, 9 (07) :1213-1222
[9]
HIERARCHICAL MIXTURES OF EXPERTS AND THE EM ALGORITHM [J].
JORDAN, MI ;
JACOBS, RA .
NEURAL COMPUTATION, 1994, 6 (02) :181-214
[10]
ROBUST STATISTICAL MODELING USING THE T-DISTRIBUTION [J].
LANGE, KL ;
LITTLE, RJA ;
TAYLOR, JMG .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1989, 84 (408) :881-896