Fast stable restricted maximum likelihood and marginal likelihood estimation of semiparametric generalized linear models

被引:5265
作者
Wood, Simon N. [1 ]
机构
[1] Univ Bath, Dept Mat Sci, Bath BA2 7AY, Avon, England
关键词
Adaptive smoothing; Generalized additive mixed model; Generalized additive model; Generalized cross-validation; Marginal likelihood; Model selection; Penalized generalized linear model; Penalized regression splines; Restricted maximum likelihood; Scalar on function regression; Stable computation; STRUCTURED ADDITIVE REGRESSION; SMOOTHING PARAMETER SELECTION; COMPONENT ESTIMATION; BAYESIAN-INFERENCE; CROSS-VALIDATION; INFORMATION; SPLINES;
D O I
10.1111/j.1467-9868.2010.00749.x
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Recent work by Reiss and Ogden provides a theoretical basis for sometimes preferring restricted maximum likelihood (REML) to generalized cross-validation (GCV) for smoothing parameter selection in semiparametric regression. However, existing REML or marginal likelihood (ML) based methods for semiparametric generalized linear models (GLMs) use iterative REML or ML estimation of the smoothing parameters of working linear approximations to the GLM. Such indirect schemes need not converge and fail to do so in a non-negligible proportion of practical analyses. By contrast, very reliable prediction error criteria smoothing parameter selection methods are available, based on direct optimization of GCV, or related criteria, for the GLM itself. Since such methods directly optimize properly defined functions of the smoothing parameters, they have much more reliable convergence properties. The paper develops the first such method for REML or ML estimation of smoothing parameters. A Laplace approximation is used to obtain an approximate REML or ML for any GLM, which is suitable for efficient direct optimization. This REML or ML criterion requires that Newton-Raphson iteration, rather than Fisher scoring, be used for GLM fitting, and a computationally stable approach to this is proposed. The REML or ML criterion itself is optimized by a Newton method, with the derivatives required obtained by a mixture of implicit differentiation and direct methods. The method will cope with numerical rank deficiency in the fitted model and in fact provides a slight improvement in numerical robustness on the earlier method of Wood for prediction error criteria based smoothness selection. Simulation results suggest that the new REML and ML methods offer some improvement in mean-square error performance relative to GCV or Akaike's information criterion in most cases, without the small number of severe undersmoothing failures to which Akaike's information criterion and GCV are prone. This is achieved at the same computational cost as GCV or Akaike's information criterion. The new approach also eliminates the convergence failures of previous REML- or ML-based approaches for penalized GLMs and usually has lower computational cost than these alternatives. Example applications are presented in adaptive smoothing, scalar on function regression and generalized additive model selection.
引用
收藏
页码:3 / 36
页数:34
相关论文
共 65 条
[1]  
Anderson E., 1999, LAPACK Users Guide, V3rd edn., DOI [10.1137/1.9780898719604, DOI 10.1137/1.9780898719604]
[2]  
ANDERSSEN RS, 1974, TECHNOMETRICS, V16, P69
[3]  
[Anonymous], 2003, Semiparametric Regression
[4]  
[Anonymous], 1994, Nonparametric Regression and Generalized Linear Models: A Roughness Penalty
[5]   APPROXIMATE INFERENCE IN GENERALIZED LINEAR MIXED MODELS [J].
BRESLOW, NE ;
CLAYTON, DG .
JOURNAL OF THE AMERICAN STATISTICAL ASSOCIATION, 1993, 88 (421) :9-25
[6]   Generalized structured additive regression based on Bayesian P-splines [J].
Brezger, A ;
Lang, S .
COMPUTATIONAL STATISTICS & DATA ANALYSIS, 2006, 50 (04) :967-991
[7]  
BREZGER A, 2007, BAYESX 1 5 0
[8]   DESCENT METHOD FOR UNIFORM SOLUTION TO OVER-DETERMINED SYSTEMS OF LINEAR EQUATIONS [J].
CLINE, AK .
SIAM JOURNAL ON NUMERICAL ANALYSIS, 1976, 13 (03) :293-309
[9]  
Craven P., 1979, Numerische Mathematik, V31, P377, DOI 10.1007/BF01404567
[10]  
Davison A.C., 2003, Statistical Models