Design of Reinforcement Learning Parameters for Seamless Application of Adaptive Traffic Signal Control

被引:101
作者
El-Tantawy, Samah [1 ,2 ]
Abdulhai, Baher [1 ]
Abdelgawad, Hossam [1 ,3 ]
机构
[1] Univ Toronto, Dept Civil Engn, Toronto, ON, Canada
[2] Cairo Univ, Dept Engn Math, Giza, Egypt
[3] Cairo Univ, Fac Engn, Dept Civil Engn, Giza 12211, Egypt
关键词
Adaptive Traffic Signal Control; Reinforcement Learning; Temporal Difference Learning; SYSTEMS; NETWORK;
D O I
10.1080/15472450.2013.810991
中图分类号
U [交通运输];
学科分类号
08 ; 0823 ;
摘要
Adaptive traffic signal control (ATSC) is a promising technique to alleviate traffic congestion. This article focuses on the development of an adaptive traffic signal control system using Reinforcement Learning (RL) as one of the efficient approaches to solve such stochastic closed loop optimal control problem. A generic RL control engine is developed and applied to a multi-phase traffic signal at an isolated intersection in Downtown Toronto in a simulation environment. Paramics, a microscopic simulation platform, is used to train and evaluate the adaptive traffic control system. This article investigates the following dimensions of the control problem: 1) RL learning methods, 2) traffic state representations, 3) action selection methods, 4) traffic signal phasing schemes, 5) reward definitions, and 6) variability of flow arrivals to the intersection. The system was tested on three networks (i.e., small, medium, large-scale) to ensure seamless transferability of the system design and results. The RL controller is benchmarked against optimized pretimed control and actuated control. The RL-based controller saves 48% average vehicle delay when compared to optimized pretimed controller and fully-actuated controller. In addition, the effect of the best design of RL-based ATSC system is tested on a large-scale application of 59 intersections in downtown Toronto and the results are compared versus the base case scenario of signal control systems in the field which are mix of pretimed and actuated controllers. The RL-based ATSC results in the following savings: average delay (27%), queue length (28%), and l CO2 emission factors (28%).
引用
收藏
页码:227 / 245
页数:19
相关论文
共 34 条
[1]   Reinforcement learning: Introduction to theory and potential for transport applications [J].
Abdulhai, B ;
Kattan, L .
CANADIAN JOURNAL OF CIVIL ENGINEERING, 2003, 30 (06) :981-991
[2]   Reinforcement learning for True Adaptive traffic signal control [J].
Abdulhai, B ;
Pringle, R ;
Karakoulas, GJ .
JOURNAL OF TRANSPORTATION ENGINEERING, 2003, 129 (03) :278-285
[3]  
[Anonymous], 2003, Simulation-Based Optimization: Parametric Optimization Tech- niques Reinforcement Learning
[4]  
[Anonymous], 1997, IEEE T AUTOM CONTROL, DOI DOI 10.1109/TAC.1997.633847
[5]   Reinforcement learning-based multi-agent system for network traffic signal control [J].
Arel, I. ;
Liu, C. ;
Urbanik, T. ;
Kohls, A. G. .
IET INTELLIGENT TRANSPORT SYSTEMS, 2010, 4 (02) :128-135
[6]   Urban traffic signal control using reinforcement learning agents [J].
Balaji, P. G. ;
German, X. ;
Srinivasan, D. .
IET INTELLIGENT TRANSPORT SYSTEMS, 2010, 4 (03) :177-188
[7]  
Barto A.G., 1998, Introduction to Reinforcement Learning, VVolume 125
[8]   Opportunities for multiagent systems and multiagent reinforcement learning in traffic control [J].
Bazzan, Ana L. C. .
AUTONOMOUS AGENTS AND MULTI-AGENT SYSTEMS, 2009, 18 (03) :342-375
[10]  
Camponogara E., 2003, 11 PORT C ART INT