A BYY scale-incremental EM algorithm for Gaussian mixture learning

被引:15
作者
Li, Lei
Ma, Jinwen [1 ]
机构
[1] Peking Univ, Sch Math Sci, Dept Informat Sci, Beijing 100871, Peoples R China
基金
中国国家自然科学基金;
关键词
Bayesian Ying-Yang (BYY) harmony learning; Gaussian mixture; EM algorithm; Model selection; Unsupervised image segmentation;
D O I
10.1016/j.amc.2008.05.076
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Gaussian mixture model has been used extensively in the fields of information processing and data analysis. However, its model selection, i.e., the selection of number of components or Gaussians in the mixture, is still a difficult problem. Fortunately, the new established Bayesian Ying-Yang (BYY) harmony function provides an efficient criterion for the model selection of Gaussian mixture with a set of sample data. In this paper, we propose a BYY scale-incremental EM algorithm for Gaussian mixture learning via a component split rule to increase the BYY harmony function incrementally. Particularly, starting from two components and adding one component sequentially via the split rule after each EM procedure until a maximum number of components, the algorithm increases the scale of the mixture incrementally and leads to the maximization of the BYY harmony function, together with the correct model selection and a good parameter estimation of the Gaussian mixture. It is demonstrated well by the simulation experiments that this BYY scale-incremental EM algorithm can make both model selection and parameter estimation efficiently for Gaussian mixture modeling. Moreover, the BYY scale-incremental EM algorithm is successfully applied to two real-life data sets, including Iris data classification and unsupervised color image segmentation. (C) 2008 Elsevier Inc. All rights reserved.
引用
收藏
页码:832 / 840
页数:9
相关论文
共 32 条
[1]   NEW LOOK AT STATISTICAL-MODEL IDENTIFICATION [J].
AKAIKE, H .
IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 1974, AC19 (06) :716-723
[2]   Generalized competitive clustering for image segmentation [J].
Boujemaa, N .
PEACHFUZZ 2000 : 19TH INTERNATIONAL CONFERENCE OF THE NORTH AMERICAN FUZZY INFORMATION PROCESSING SOCIETY - NAFIPS, 2000, :133-137
[4]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[5]   BAYESIAN DENSITY-ESTIMATION AND INFERENCE USING MIXTURES [J].
ESCOBAR, MD ;
WEST, M .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1995, 90 (430) :577-588
[6]   SUBCELLULAR-LOCALIZATION OF STEROL SYNTHESIZING ENZYMES IN MAIZE COLEOPTILES [J].
HARTMANN, MA ;
FONTENEAU, P ;
BENVENISTE, P .
PLANT SCIENCE LETTERS, 1977, 8 (01) :45-51
[7]  
MA J, 2006, P 2006 INT JOINT C N, P8255
[8]   A fast fixed-point BYY harmony learning algorithm on Gaussian mixture with automated model selection [J].
Ma, Jinwen ;
He, Xuefeng .
PATTERN RECOGNITION LETTERS, 2008, 29 (06) :701-711
[9]   BYY harmony learning on finite mixture: Adaptive gradient implementation and a floating RPCL mechanism [J].
Ma, Jinwen ;
Wang, Le .
NEURAL PROCESSING LETTERS, 2006, 24 (01) :19-40
[10]   Asymptotic convergence rate of the EM algorithm for Gaussian mixtures [J].
Ma, JW ;
Xu, L ;
Jordan, MI .
NEURAL COMPUTATION, 2000, 12 (12) :2881-2907