OPTIMAL INFINITE-HORIZON UNDISCOUNTED CONTROL OF FINITE PROBABILISTIC SYSTEMS

被引:42
作者
PLATZMAN, LK [1 ]
机构
[1] BELL TEL LABS INC,NAPERVILLE,IL 60540
关键词
D O I
10.1137/0318028
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A finite-point, finite-state, finite-output stochastic control problem with imperfect state observation and-classical information pattern is shown to be meaningful as the horizon increases without bound and the discount rate approaches unity. The plant model, a finite probabilistic system, includes the Markov decision and partially-observed Markov decision problems as special cases. Under conditions resembling controllability and observability in linear systems it is shown that: an optimal strategy exists, it may be realized by a stationary policy on the state estimate, its performance does not depend on the initial state distribution, and convergence rates for its finite-horizon and discounted performances are readily established.
引用
收藏
页码:362 / 380
页数:19
相关论文
共 33 条
[11]  
DRAKE A, 1962, THESIS MASSACHUSETTS
[12]  
Feldbaum A., 1965, OPTIMAL CONTROL SYST
[13]   CONDITIONS FOR EQUIVALENCE OF OPTIMALITY CRITERIA IN DYNAMIC-PROGRAMMING [J].
FLYNN, J .
ANNALS OF STATISTICS, 1976, 4 (05) :936-953
[14]  
Hastings N.A.J., 1973, DYNAMIC PROGRAMMING
[15]  
Howard R.A., 1971, DYNAMIC PROBABILISTI, V1
[16]  
Howard R.A., DYNAMIC PROBABILISTI, VII
[17]  
Howard RonaldA., 1960, DYNAMIC PROGRAMMING
[18]  
Kalman R. E., 1969, TOPICS MATH SYSTEM T
[19]  
Kushner Harold, 1971, INTRO STOCHASTIC CON
[20]  
Mine H., 1970, MARKOVIAN DECISION P