Effectiveness of Q-learning as a tool for calibrating agent-based supply network models

被引:24
作者
Zhang, Y. [1 ]
Bhattacharyya, S. [2 ]
机构
[1] Univ Illinois, Dept Management Informat Syst, Coll Business & Management, Springfield, IL 62703 USA
[2] Univ Illinois, Dept Informat & Decis Sci, Chicago, IL 60607 USA
关键词
Agent-based modelling; Q-learning; Supply chain management (SCM); Computational techniques; Distributed artificial intelligence (DAI);
D O I
10.1080/17517570701275390
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
This paper examines effectiveness of Q-learning as a tool for specifying agent attributes and behaviours in agent-based supply network models. Agent-based modelling (ABM) has been increasingly employed to study supply chain and supply network problems. A challenging task in building agent-based supply network models is to properly specify agent attributes and behaviours. Machine learning techniques, such as Q-learning, can be a useful tool for this purpose. Q-learning is a reinforcement learning technique that has been shown to be an effective adaptation and searching mechanism in distributed settings. In this study, Q-learning is employed by supply network agents to search for 'optimal' values for a parameter in their operating policies simultaneously and independently. Methods are designed to identify the 'optimal' parameter values against which effectiveness of the learning is evaluated. Robustness of the learning's effectiveness is also examined through consideration of different model settings and scenarios. Results show that Q-learning is very effective in finding the 'optimal' parameter values in all model settings and scenarios considered.
引用
收藏
页码:217 / 233
页数:17
相关论文
共 18 条
[1]  
[Anonymous], THESIS KINGS COLL
[2]  
Axtell R., 2000, 17 CTR SOC EC DYN BR
[3]   Performance analysis of conjoined supply chains [J].
Beamon, BM ;
Chen, VCP .
INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2001, 39 (14) :3195-3218
[4]   Pricing and replenishment strategies in a distribution system with competing retailers [J].
Bernstein, F ;
Federgruen, A .
OPERATIONS RESEARCH, 2003, 51 (03) :409-426
[5]  
Bruhn P., 2000, EC SIMULATIONS SWARM, P251
[6]   OPTIMAL POLICIES FOR A MULTI-ECHELON INVENTORY PROBLEM [J].
CLARK, AJ ;
SCARF, H .
MANAGEMENT SCIENCE, 1960, 6 (04) :475-490
[7]   Learning Curve: A Simulation-Based Approach to Dynamic Pricing [J].
Joan Morris DiMicco ;
Pattie Maes ;
Amy Greenwald .
Electronic Commerce Research, 2003, 3 (3-4) :245-276
[8]  
GREENWALD A, 1999, P 1 ACM C EL COMM
[9]   Learning competitive pricing strategies by multi-agent reinforcement learning [J].
Kutschinski, E ;
Uthmann, T ;
Polani, D .
JOURNAL OF ECONOMIC DYNAMICS & CONTROL, 2003, 27 (11-12) :2207-2218
[10]  
Lin F., 2000, EC SIMULATIONS SWARM, P225