Stochastic gradient boosting

被引:4674
作者
Friedman, JH
机构
[1] Stanford Univ, Dept Stat, Stanford, CA 94305 USA
[2] Stanford Univ, Stanford Linear Accelerator Ctr, Stanford, CA 94305 USA
基金
美国国家科学基金会;
关键词
Data reduction - Least squares approximations - Regression analysis - Robustness (control systems) - Stochastic control systems;
D O I
10.1016/S0167-9473(01)00065-2
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Gradient boosting constructs additive regression models by sequentially fitting a simple parameterized function (base learner) to current "pseudo"-residuals by least squares at each iteration. The pseudo-residuals are the gradient of the loss functional being minimized, with respect to the model values at each training data point evaluated at the current step. It is shown that both the approximation accuracy and execution speed of gradient boosting can be substantially improved by incorporating randomization into the procedure. Specifically, at each iteration a subsample of the training data is drawn at random (without replacement) from the full training data set. This randomly selected subsample is then used in place of the full sample to fit the base learner and compute the model update for the current iteration. This randomized approach also increases robustness against overcapacity of the base learner. (C) 2002 Elsevier Science B.V. All rights reserved.
引用
收藏
页码:367 / 378
页数:12
相关论文
共 5 条
  • [1] [Anonymous], 1999, USING ADAPTIVE BAGGI
  • [2] Bagging predictors
    Breiman, L
    [J]. MACHINE LEARNING, 1996, 24 (02) : 123 - 140
  • [3] Freund Y, 1996, Experiments with a new boosting algorithm. In proceedings 13th Int Conf Mach learn. Pp.148-156, P45
  • [4] Friedman J, Towards an expansive epistemology: Norms, action and the social sphere
  • [5] Friedman J., 1998, ADDITIVE LOGISTIC RE