Improve multi-instance neural networks through feature selection

被引:76
作者
Zhang, ML [1 ]
Zhou, ZH [1 ]
机构
[1] Nanjing Univ, State Key Lab Novel Software Technol, Nanjing 210093, Peoples R China
基金
中国国家自然科学基金;
关键词
backpropagation; feature selection; machine learning; multi-instance learning; neural networks;
D O I
10.1023/B:NEPL.0000016836.03614.9f
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
bMulti-instance learning is regarded as a new learning framework where the training examples are bags composed of instances without labels, and the task is to predict the labels of unseen bags through analyzing the training bags with known labels. Recently, a multi-instance neural network BP-MIP was proposed. In this paper, BP-MIP is improved through adopting two different feature selection techniques, i.e. feature scaling with Diverse Density and feature reduction with principal component analysis. In detail, before feature vectors are fed to a BP-MIP neural network, they are scaled by the feature weights found by running Diverse Density on the training data, or projected by a linear transformation matrix formed by principal component analysis. Experiments show that these feature selection mechanisms can significantly improve the performance of BP-MIP.
引用
收藏
页码:1 / 10
页数:10
相关论文
共 19 条
  • [11] Pearl J., 1989, Probabilistic reasoning in intelligent systems: networks of plausible inference, DOI DOI 10.1016/C2009-0-27609-4
  • [12] Ray S., 2001, Proceedings of the Eighteenth International Conference on Machine Learning, ICML'01, P425
  • [13] RUFFO G, 2000, THESIS U TURIN TORIN
  • [14] Rumelhart D.E., 1987, Parallel Distributed Processing: Explorations in the Microstructure of Cognition, P318
  • [15] Wang J., 2000, PROC 17 INT C MACHIN, P1119
  • [16] Yang C., 2000, Proceedings of 16th International Conference on Data Engineering (Cat. No.00CB37073), P233, DOI 10.1109/ICDE.2000.839416
  • [17] Zhang Q, 2002, ADV NEUR IN, V14, P1073
  • [18] Zhou ZH, 2003, LECT NOTES ARTIF INT, V2837, P492
  • [19] ZHOU ZH, 2002, NEURAL NETWORKS MULT