Random Subspace Ensembles for fMRI Classification

被引:172
作者
Kuncheva, Ludmila I. [1 ]
Rodriguez, Juan J. [2 ]
Plumpton, Catrin O. [1 ]
Linden, David E. J. [3 ,4 ]
Johnston, Stephen J. [5 ]
机构
[1] Bangor Univ, Sch Comp Sci, Bangor LL57 1UT, Gwynedd, Wales
[2] Univ Burgos, Dept Ingn Civil, Burgos 09006, Spain
[3] Bangor Univ, Bangor Imaging Unit, Wolfson Ctr Clin & Cognit Neurosci, Sch Psychol, Bangor LL57 2AS, Gwynedd, Wales
[4] NW Wales NHS Trust, Bangor LL57 2PW, Gwynedd, Wales
[5] Bangor Univ, Bangor Imaging Unit, Wolfson Ctr Clin & Cognit Neurosci, Sch Psychol, Bangor LL57 1UT, Gwynedd, Wales
基金
英国医学研究理事会;
关键词
Classifier ensembles; functional magnetic resonance imaging (fMRI) data analysis; multivariate methods; pattern recognition; random subspace (RS) method; SUPPORT VECTOR MACHINES; BRAIN ACTIVITY; SELECTION; SVM; PATTERNS;
D O I
10.1109/TMI.2009.2037756
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Classification of brain images obtained through functional magnetic resonance imaging (fMRI) poses a serious challenge to pattern recognition and machine learning due to the extremely large feature-to-instance ratio. This calls for revision and adaptation of the current state-of-the-art classification methods. We investigate the suitability of the random subspace (RS) ensemble method for fMRI classification. RS samples from the original feature set and builds one (base) classifier on each subset. The ensemble assigns a class label by either majority voting or averaging of output probabilities. Looking for guidelines for setting the two parameters of the method-ensemble size and feature sample size-we introduce three criteria calculated through these parameters: usability of the selected feature sets, coverage of the set of "important" features, and feature set diversity. Optimized together, these criteria work toward producing accurate and diverse individual classifiers. RS was tested on three fMRI datasets from single-subject experiments: the Haxby et al. data (Haxby, 2001.) and two datasets collected in-house. We found that RS with support vector machines (SVM) as the base classifier outperformed single classifiers as well as some of the most widely used classifier ensembles such as bagging, AdaBoost, random forest, and rotation forest. The closest rivals were the single SVM and bagging of SVM classifiers. We use kappa-error diagrams to understand the success of RS.
引用
收藏
页码:531 / 542
页数:12
相关论文
共 57 条
  • [1] [Anonymous], 1998, Ph.D. Thesis
  • [2] [Anonymous], 2001, Pattern Classification
  • [3] Feature selection combined with random subspace ensemble for gene expression based diagnosis of malignancies
    Bertoni, Alberto
    Folgieri, Raffaella
    Valentini, Giorgio
    [J]. Biological and Artificial Intelligence Environments, 2005, : 29 - 35
  • [4] Bishop CM., 1995, NEURAL NETWORKS PATT
  • [5] Random forests
    Breiman, L
    [J]. MACHINE LEARNING, 2001, 45 (01) : 5 - 32
  • [6] Breiman L, 1996, MACH LEARN, V24, P123, DOI 10.1023/A:1018054314350
  • [7] Brown G., 2005, Information Fusion, V6, P5, DOI 10.1016/j.inffus.2004.04.004
  • [8] Brown G., 2009, ENCY MACHINE LEARNIN
  • [9] A tutorial on Support Vector Machines for pattern recognition
    Burges, CJC
    [J]. DATA MINING AND KNOWLEDGE DISCOVERY, 1998, 2 (02) : 121 - 167
  • [10] Local pattern classification differentiates processes of economic valuation
    Clithero, John A.
    Carter, R. McKell
    Huettel, Scott A.
    [J]. NEUROIMAGE, 2009, 45 (04) : 1329 - 1338