Computing normalizing constants for finite mixture models via incremental mixture importance sampling (IMIS)

被引:32
作者
Steele, Russell J.
Raftery, Adrian E.
Emond, Mary J.
机构
[1] McGill Univ, Montreal, PQ H3A 2K6, Canada
[2] Univ Washington, Seattle, WA 98195 USA
关键词
Bayes factor; Bayesian model averaging; dirichlet-multinomial; defensive mixture importance sampling; Gibbs sampling; label-switching; Markov chain Monte Carlo; multimodality;
D O I
10.1198/106186006X132358
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
This article proposes a method for approximating integrated likelihoods in finite mixture models. We formulate the model in terms of the unobserved group memberships, z, and make them the variables of integration. The integral is then evaluated using importance sampling over the z. We propose an adaptive importance sampling function which is itself a mixture, with two types of component distributions, one concentrated and one diffuse. The more concentrated type of component serves the usual purpose of an importance sampling function, sampling mostly group assignments of high posterior probability. The less concentrated type of component allows for the importance sampling function to explore the space in a controlled way to find other, unvisited assignments with high posterior probability. Components are added adaptively, one at a time, to cover areas of high posterior probability not well covered by the current importance sampling function. The method is called incremental mixture importance sampling (IMIS). IMIS is easy to implement and to monitor for convergence. It scales easily for higher dimensional mixture distributions when a conjugate prior is specified for the mixture parameters. The simulated values on which the estimate is based are independent, which allows for straightforward estimation of standard errors. The self-monitoring aspects of the method make it easier to adjust tuning parameters in the course of estimation than standard Markov chain Monte Carlo algorithms. With only small modifications to the code, one can use the method for a wide variety of mixture distributions of different dimensions. The method performed well in simulations and in mixture problems in astronomy and medical research.
引用
收藏
页码:712 / 734
页数:23
相关论文
共 52 条
[1]  
[Anonymous], 1979, Monte Carlo Methods, DOI DOI 10.1007/978-94-009-5819-7
[2]   COMPUTATIONAL ASPECTS OF FITTING A MIXTURE OF 2 NORMAL-DISTRIBUTIONS USING MAXIMUM-LIKELIHOOD [J].
ATWOOD, LD ;
WILSON, AF ;
ELSTON, RC ;
BAILEYWILSON, JE .
COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 1992, 21 (03) :769-781
[3]   Assessing a mixture model for clustering with the integrated completed likelihood [J].
Biernacki, C ;
Celeux, G ;
Govaert, G .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2000, 22 (07) :719-725
[4]   Computational and inferential difficulties with mixture posterior distributions. [J].
Celeux, G ;
Hurn, M ;
Robert, CP .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2000, 95 (451) :957-970
[5]  
CELEUX G, 1997, J R STAT SOC, V59, P775
[6]  
CHEESEMAN P, 1995, ADV KNOWLEDGE DISCOV, P153
[7]  
Chen M. H., 2000, MONTE CARLO METHODS
[8]   Marginal likelihood from the Gibbs output [J].
Chib, S .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1995, 90 (432) :1313-1321
[9]   Marginal likelihood from the Metropolis-Hastings output [J].
Chib, S ;
Jeliazkov, I .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (453) :270-281
[10]   A new mixture model approach to analyzing allelic-loss data using Bayes factors [J].
Desai, M ;
Emond, MJ .
BMC BIOINFORMATICS, 2004, 5 (1)