Criticality of predictors in multiple regression

被引:33
作者
Azen, R
Budescu, DV
Reiser, B
机构
[1] Univ Wisconsin, Dept Educ Psychol, Milwaukee, WI 53201 USA
[2] Univ Illinois, Dept Psychol, Urbana, IL 61801 USA
[3] Univ Haifa, Dept Stat, IL-31999 Haifa, Israel
关键词
D O I
10.1348/000711001159483
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
A new method is proposed for comparing all predictors in a multiple regression model. This method generates a measure of predictor criticality, which is distinct from and has several advantages over traditional indices of predictor importance. Using the bootstrapping (resampling with replacement) procedure, a large number of samples are obtained from a given data set which contains one response variable and p predictors. For each sample, all 2(p) - 1 subset regression models are fitted and the best subset model is selected. Thus, the (multinomial) distribution of the probability that each of the 2(p) - 1 subsets is 'the best' model for the data set is obtained. A predictor's criticality is defined as a function of the probabilities associated with the models that include the predictor. That is, a predictor which is included in a large number of probable models is critical to the identification of the best-fitting regression model and, therefore, to the prediction of the response variable. The procedure can be applied to fixed and random regression models and can use any measure of goodness of fit (e.g., adjusted R-2, C-p, AIC) for identifying the best model. Several criticality measures can be defined by using different combinations of the probabilities of the best-fitting models, and asymptotic confidence intervals for each variable's criticality can be derived. The procedure is illustrated with several examples.
引用
收藏
页码:201 / 225
页数:25
相关论文
共 23 条