Hierarchically organized behavior and its neural foundations: A reinforcement learning perspective

被引:381
作者
Botvinick, Matthew M. [1 ]
Niv, Yael [1 ]
Barto, Andrew C. [2 ]
机构
[1] Princeton Univ, Princeton Neurosci Inst, Dept Psychol, Princeton, NJ 08540 USA
[2] Univ Massachusetts, Dept Comp Sci, Amherst, MA 01003 USA
关键词
Reinforcement learning; Prefrontal cortex; TONICALLY ACTIVE NEURONS; PREFRONTAL CORTEX; BASAL-GANGLIA; ORBITOFRONTAL CORTEX; COGNITIVE CONTROL; WORKING-MEMORY; TEMPORAL ORGANIZATION; SERIAL ORDER; DOPAMINE; PERCEPTION;
D O I
10.1016/j.cognition.2008.08.011
中图分类号
B84 [心理学];
学科分类号
04 ; 0402 ;
摘要
Research on human and animal behavior has long emphasized its hierarchical structure the divisibility of ongoing behavior into discrete tasks, which are comprised of subtask sequences, which in turn are built of simple actions. The hierarchical structure of behavior has also been of enduring interest within neuroscience, where it has been widely considered to reflect prefrontal cortical functions. In this paper, we reexamine behavioral hierarchy and its neural substrates from the point of view of recent developments in computational reinforcement learning. Specifically, we consider a set of approaches known collectively as hierarchical reinforcement learning, which extend the reinforcement learning paradigm by allowing the learning agent to aggregate actions into reusable subroutines or skills. A close took at the components of hierarchical reinforcement learning suggests how they might map onto neural structures, in particular regions within the dorsolateral and orbital prefrontal cortex. It also suggests specific ways in which hierarchical reinforcement learning might provide a complement to existing psychological models of hierarchically structured behavior. A particularly important question that hierarchical reinforcement learning brings to the fore is that of how learning identifies new action routines that are likely to provide useful building blocks in solving a wide range of future problems. Here and at many other points. hierarchical reinforcement learning offers an appealing framework for investigating the computational and neural underpinnings of hierarchically structured behavior. (C) 2008 Elsevier B.V. All rights reserved.
引用
收藏
页码:262 / 280
页数:19
相关论文
共 175 条
[71]   LEARNING MOTIVATED BY A MANIPULATION DRIVE [J].
HARLOW, HF ;
HARLOW, MK ;
MEYER, DR .
JOURNAL OF EXPERIMENTAL PSYCHOLOGY, 1950, 40 (02) :228-234
[72]   Heterarchical reinforcement-learning model for integration of multiple cortico-striatal loops: fMRI examination in stimulus-action-reward association learning [J].
Haruno, Masahiko ;
Kawato, Mitsuo .
NEURAL NETWORKS, 2006, 19 (08) :1242-1254
[73]  
HAYES-ROTH B, 1979, Cognitive Science, V3, P275, DOI 10.1016/S0364-0213(79)80010-5
[74]  
Holroyd CB, 2002, PSYCHOL REV, V109, P679, DOI [10.1037//0033-295X.109.4.679, 10.1037/0033-295X.109.4.679]
[75]   Task-dependent selectivity of movement-related neuronal activity in the primate prefrontal cortex [J].
Hoshi, E ;
Shima, K ;
Tanji, J .
JOURNAL OF NEUROPHYSIOLOGY, 1998, 80 (06) :3392-3397
[76]  
Houk J. C., 1995, Models of Information Processing in the Basal Ganglia, P249
[77]   Actor-critic models of the basal ganglia: new anatomical and computational perspectives [J].
Joel, D ;
Niv, Y ;
Ruppin, E .
NEURAL NETWORKS, 2002, 15 (4-6) :535-547
[78]   Neural activity in monkey prefrontal cortex is modulated by task context and behavioral instruction during delayed-match-to-sample and conditional prosaccade-antisaccade tasks [J].
Johnston, Kevin ;
Everling, Stefan .
JOURNAL OF COGNITIVE NEUROSCIENCE, 2006, 18 (05) :749-765
[79]  
Jonsson A, 2001, ADV NEUR IN, V13, P1054
[80]  
JONSSON A, 2005, P INT C MACH LEARN, V22