A WEAKLY INFORMATIVE DEFAULT PRIOR DISTRIBUTION FOR LOGISTIC AND OTHER REGRESSION MODELS

被引:1424
作者
Gelman, Andrew [1 ,2 ]
Jakulin, Aleks [1 ]
Pittau, Maria Grazia [3 ]
Su, Yu-Sung [2 ]
机构
[1] Columbia Univ, Dept Stat, New York, NY 10027 USA
[2] Columbia Univ, Dept Polit Sci, New York, NY 10027 USA
[3] Univ Rome, Dept Econ, Rome, Italy
基金
美国国家卫生研究院; 美国国家科学基金会;
关键词
Bayesian inference; generalized linear model; least squares; hierarchical model; linear regression; logistic regression; multilevel model; noninformative prior distribution; weakly informative prior distribution;
D O I
10.1214/08-AOAS191
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We propose a new prior distribution for classical (nonhierarchical) logistic regression models, constructed by first scaling all nonbinary variables to have mean 0 and standard deviation 0.5, and then placing independent Student-t prior distributions on the coefficients. As a default choice, we recommend the Cauchy distribution with center 0 and scale 2.5, which in the simplest setting is a longer-tailed version of the distribution attained by assuming one-half additional success and one-half additional failure in a logistic regression. Cross-validation on a corpus of datasets shows the Cauchy class of prior distributions to outperform existing implementations of Gaussian and Laplace priors. We recommend this prior distribution as a default choice for routine applied use. It has the advantage of always giving answers, even when there is complete separation in logistic regression (a common problem, even when the sample size is large and the number of predictors is small), and also automatically applying more shrinkage to higher-order interactions. This can be useful in routine data analysis as well as in automated procedures such as chained equations for missing-data imputation. We implement a procedure to fit generalized linear models in R with the Student-t prior distribution by incorporating an approximate EM algorithm into the usual iteratively weighted least squares. We illustrate with several applications, including a series of logistic regressions predicting voting preferences, a small bioassay experiment, and an imputation model for a public health data set.
引用
收藏
页码:1360 / 1383
页数:24
相关论文
共 50 条
[1]   Approximate is better than "exact" for interval estimation of binomial proportions [J].
Agresti, A ;
Coull, BA .
AMERICAN STATISTICIAN, 1998, 52 (02) :119-126
[2]  
ALBERT A, 1984, BIOMETRIKA, V71, P1
[3]  
Asuncion A., 2007, UCI MACHINE LEARNING
[4]   A new perspective on priors for generalized linear models [J].
Bedrick, EJ ;
Christensen, R ;
Johnson, W .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1996, 91 (436) :1450-1460
[5]   ROBUST BAYES AND EMPIRICAL BAYES ANALYSIS WITH EPSILON-CONTAMINATED PRIORS [J].
BERGER, J ;
BERLINER, LM .
ANNALS OF STATISTICS, 1986, 14 (02) :461-486
[6]  
BERNARDO JM, 1979, J R STAT SOC B, V41, P113
[7]  
Brier G. W., 1950, MON WEATHER REV, V78, P1, DOI [10.1175/1520-0493(1950)078%3C0001:VOFEIT%3E2.0.CO
[8]  
2, 10.1175/1520-0493(1950)0782.0.co
[9]  
2, DOI 10.1016/0016-0032(94)90228-3]
[10]  
C Liu, 2004, APPL BAYESIAN MODELI, P227