EXISTENCE OF OPTIMAL STATIONARY POLICIES IN AVERAGE REWARD MARKOV DECISION-PROCESSES WITH A RECURRENT STATE

被引:3
作者
CAVAZOSCADENA, R [1 ]
机构
[1] TEXAS TECH UNIV,DEPT MATH,LUBBOCK,TX 79409
关键词
AVERAGE REWARD CRITERIA; OPTIMAL STATIONARY POLICIES; RECURRENT STATE; RENEWAL PROCESSES;
D O I
10.1007/BF01189029
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider discrete-time average reward Markov decision processes with denumerable state space and bounded reward function. Under structural restrictions on the model the existence of an optimal stationary policy is proved; both the lim inf and lim sup average criteria are considered. In contrast to the usual approach our results do not rely on the average regard optimality equation. Rather, the arguments are based on well-known facts from Renewal Theory.
引用
收藏
页码:171 / 194
页数:24
相关论文
共 15 条
[1]  
Ash R.B., 1972, PROBABILITY MATH STA, V11
[2]  
BARAS JS, 1984, SRR8417 U MAR EL ENG
[4]   NECESSARY AND SUFFICIENT CONDITIONS FOR A BOUNDED SOLUTION TO THE OPTIMALITY EQUATION IN AVERAGE REWARD MARKOV DECISION CHAINS [J].
CAVAZOSCADENA, R .
SYSTEMS & CONTROL LETTERS, 1988, 10 (01) :71-78
[5]   CONDITIONS FOR EQUIVALENCE OF OPTIMALITY CRITERIA IN DYNAMIC-PROGRAMMING [J].
FLYNN, J .
ANNALS OF STATISTICS, 1976, 4 (05) :936-953
[6]  
HEYMAN DP, 1984, STOCHASTIC MODELS OP, V2
[7]  
HINDERER K, 1970, F NONSTATIONARY DYNA
[8]  
Munkres J.R., 1975, TOPOLOGY 1 COURSE
[9]  
Ross, 1983, STOCHASTIC PROCESSES
[10]  
Ross S., 1983, INTRO STOCHASTIC DYN