An analysis of temporal-difference learning with function approximation

被引:651
作者
Tsitsiklis, JN
VanRoy, B
机构
[1] Laboratory for Information and Decision Systems, Massachusetts Institute of Technology, Cambridge
基金
美国国家科学基金会;
关键词
dynamic programming; function approximation; Markov chains; neuro-dynamic programming; reinforcement learning; temporal-difference learning;
D O I
10.1109/9.580874
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We discuss the temporal-difference learning algorithm, as applied to approximating the cost-to-go function of an infinite-horizon discounted Markov chain, The algorithm we analyze updates parameters of a linear function approximator online during a single endless trajectory of an irreducible aperiodic Markov chain with a finite or infinite state space, We present a proof of convergence (with probability one), a characterization of the limit of convergence, and a bound on the resulting approximation error, Furthermore, our analysis is based on a new line of reasoning that provides new intuition about the dynamics of temporal difference learning. In addition to proving new and stronger positive results than those previously available, we identify the significance of online updating and potential hazards associated with the use of nonlinear function approximators, First, we prove that divergence may occur when updates are not based on trajectories of the Markov chain, This bet reconciles positive and negative results that have been discussed in the literature, regarding the soundness of temporal-difference learning, Second, we present an example illustrating the possibility of divergence when temporal-difference learning is used in the presence of a nonlinear function approximator.
引用
收藏
页码:674 / 690
页数:17
相关论文
共 25 条