Reinforcement Learning Trees

被引:112
作者
Zhu, Ruoqing [1 ]
Zeng, Donglin [1 ]
Kosorok, Michael R. [1 ]
机构
[1] Univ N Carolina, Dept Biostat, CB 7420, Chapel Hill, NC 27599 USA
关键词
Consistency; Error bound; Random forests; Reinforcement learning; Trees; RANDOM FORESTS; RANDOMIZED TREES; CLASSIFICATION; SELECTION; CONSISTENCY; LASSO; MODEL;
D O I
10.1080/01621459.2015.1036994
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
In this article, we introduce a new type of tree-based method, reinforcement learning trees (RLT), which exhibits significantly improved performance over traditional methods such as random forests (Breiman 2001) under high-dimensional settings. The innovations are threefold. First, the new method implements reinforcement learning at each selection of a splitting variable during the tree construction processes. By splitting on the variable that brings the greatest future improvement in later splits, rather than choosing the one with largest marginal effect from the immediate split, the constructed tree uses the available samples in a more efficient way. Moreover, such an approach enables linear combination cuts at little extra computational cost. Second, we propose a variable muting procedure that progressively eliminates noise variables during the construction of each individual tree. The muting procedure also takes advantage of reinforcement learning and prevents noise variables from being considered in the search for splitting rules, so that toward terminal nodes, where the sample size is small, the splitting rules are still constructed from only strong variables. Last, we investigate asymptotic properties of the proposed method under basic assumptions and discuss rationale in general settings. Supplementary materials for this article are available online.
引用
收藏
页码:1770 / 1784
页数:15
相关论文
共 28 条
  • [1] Shape quantization and recognition with randomized trees
    Amit, Y
    Geman, D
    [J]. NEURAL COMPUTATION, 1997, 9 (07) : 1545 - 1588
  • [2] [Anonymous], 2001, Computing Science and Statistics
  • [3] [Anonymous], 2001, MACH LEARN, DOI DOI 10.1023/A:1010933404324
  • [4] [Anonymous], 1984, OLSHEN STONE CLASSIF, DOI 10.2307/2530946
  • [5] Biau G, 2012, J MACH LEARN RES, V13, P1063
  • [6] Biau G, 2008, J MACH LEARN RES, V9, P2015
  • [7] Breiman L, 1996, MACH LEARN, V24, P123, DOI 10.1023/A:1018054314350
  • [8] Breiman L., 2000, Some infinite theory for predictor ensembles
  • [9] Identifying SNPs predictive of phenotype using random forests
    Bureau, A
    Dupuis, J
    Falls, K
    Lunetta, KL
    Hayward, B
    Keith, TP
    Van Eerdewegh, P
    [J]. GENETIC EPIDEMIOLOGY, 2005, 28 (02) : 171 - 182
  • [10] BART: BAYESIAN ADDITIVE REGRESSION TREES
    Chipman, Hugh A.
    George, Edward I.
    McCulloch, Robert E.
    [J]. ANNALS OF APPLIED STATISTICS, 2010, 4 (01) : 266 - 298