Ascent-based Monte Carlo expectation-maximization

被引:83
作者
Caffo, BS
Jank, W
Jones, GL
机构
[1] Johns Hopkins Univ, Dept Biostat, Baltimore, MD 21205 USA
[2] Univ Maryland, College Pk, MD 20742 USA
[3] Univ Minnesota, Minneapolis, MN USA
关键词
EM algorithm; empirical Bayes estimates; generalized linear mixed models; importance sampling; Markov chain; Monte Carlo methods;
D O I
10.1111/j.1467-9868.2005.00499.x
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
The expectation-maximization (EM) algorithm is a popular tool for maximizing likelihood functions in the presence of missing data. Unfortunately, EM often requires the evaluation of analytically intractable and high dimensional integrals. The Monte Carlo EM (MCEM) algorithm is the natural extension of EM that employs Monte Carlo methods to estimate the relevant integrals. Typically, a very large Monte Carlo sample size is required to estimate these integrals within an acceptable tolerance when the algorithm is near convergence. Even if this sample size were known at the onset of implementation of MCEM, its use throughout all iterations is wasteful, especially when accurate starting values are not available. We propose a data-driven strategy for controlling Monte Carlo resources in MCEM. The algorithm proposed improves on similar existing methods by recovering EM's ascent (i.e. likelihood increasing) property with high probability, being more robust to the effect of user-defined inputs and handling classical Monte Carlo and Markov chain Monte Carlo methods within a common framework. Because of the first of these properties we refer to the algorithm as 'ascent-based MCEM'. We apply ascent-based MCEM to a variety of examples, including one where it is used to accelerate the convergence of deterministic EM dramatically.
引用
收藏
页码:235 / 251
页数:17
相关论文
共 32 条
[1]  
Billingsley P., 1986, PROBABILITY MEASURE
[2]   A survey of Monte Carlo algorithms for maximizing the likelihood of a two-stage hierarchical model [J].
Booth, James G. ;
Hobert, James P. ;
Jank, Wolfgang .
STATISTICAL MODELLING, 2001, 1 (04) :333-349
[3]   Maximizing generalized linear mixed model likelihoods with an automated Monte Carlo EM algorithm [J].
Booth, JG ;
Hobert, JP .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-STATISTICAL METHODOLOGY, 1999, 61 :265-285
[4]   Empirical supremum rejection sampling [J].
Caffo, BS ;
Booth, JG ;
Davison, AC .
BIOMETRIKA, 2002, 89 (04) :745-754
[5]  
Christensen O.F., 2002, R NEWS, V2, P26
[6]   MAXIMUM LIKELIHOOD FROM INCOMPLETE DATA VIA EM ALGORITHM [J].
DEMPSTER, AP ;
LAIRD, NM ;
RUBIN, DB .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES B-METHODOLOGICAL, 1977, 39 (01) :1-38
[7]   Model-based geostatistics [J].
Diggle, PJ ;
Tawn, JA ;
Moyeed, RA .
JOURNAL OF THE ROYAL STATISTICAL SOCIETY SERIES C-APPLIED STATISTICS, 1998, 47 :299-326
[8]  
GEYER CJ, 1994, J ROY STAT SOC B MET, V56, P261
[9]   A correlated probit model for joint modeling of clustered binary and continuous responses [J].
Gueorguieva, RV ;
Agresti, A .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 2001, 96 (455) :1102-1112
[10]   Geometric ergodicity of Gibbs and block Gibbs samplers for a hierarchical random effects model [J].
Hobert, JP .
JOURNAL OF MULTIVARIATE ANALYSIS, 1998, 67 (02) :414-430