Active learning with statistical models

被引:1023
作者
Cohn, DA
Ghahramani, Z
Jordan, MI
机构
[1] Ctr. for Biol./Compl. Learning, Dept. of Brain Sciences, Massachusetts Inst. of Technology, Cambridge
关键词
D O I
10.1613/jair.295
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For many types of machine learning algorithms, one can compute the statistically ''optimal'' way to select training data. In this paper, we review how optimal data selection techniques have been used with feedforward neural networks. We then show how the same principles may be used to select data for two alternative, statistically-based learning architectures: mixtures of Gaussians and locally weighted regression. While the techniques for neural networks are computationally expensive and approximate, the techniques for mixtures of Gaussians and locally weighted regression are both efficient and accurate. Empirically, we observe that the optimality criterion sharply decreases the number of training examples the learner needs in order to achieve good performance.
引用
收藏
页码:129 / 145
页数:17
相关论文
共 28 条
  • [1] Angluin D., 1988, Machine Learning, V2, P319, DOI 10.1007/BF00116828
  • [2] BAUM E, 1991, IEEE T NEURAL NETWOR, V2
  • [3] Box G. E. P., 1987, Empirical model-building and response surfaces
  • [4] Cheeseman P., 1988, AAAI 88. Seventh National Conference on Artificial Intelligence, P607
  • [5] REGRESSION BY LOCAL FITTING - METHODS, PROPERTIES, AND COMPUTATIONAL ALGORITHMS
    CLEVELAND, WS
    DEVLIN, SJ
    GROSSE, E
    [J]. JOURNAL OF ECONOMETRICS, 1988, 37 (01) : 87 - 114
  • [6] COHN D, 1994, MACH LEARN, V15, P201, DOI 10.1007/BF00993277
  • [7] COHN D, 1994, ADV NEURAL INFORMATI, V6
  • [8] COHN D, 1990, ADV NEURAL INFORMATI, V2
  • [9] COHN D, 1995, AIM1552 MIT
  • [10] MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM
    DEMPSTER, AP
    LAIRD, NM
    RUBIN, DB
    [J]. JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01): : 1 - 38