LEARNING INDIRECT ACTIONS IN COMPLEX DOMAINS: ACTION SUGGESTIONS FOR AIR TRAFFIC CONTROL

被引:9
作者
Agogino, Adrian [1 ]
Tumer, Kagan [2 ]
机构
[1] NASA, Ames Res Ctr, UCSC, Moffett Field, CA 94035 USA
[2] Oregon State Univ, Corvallis, OR 97331 USA
来源
ADVANCES IN COMPLEX SYSTEMS | 2009年 / 12卷 / 4-5期
基金
美国国家科学基金会;
关键词
Air traffic control; learning indirect actions; suggestion agents; multiagent learning; CONFLICT-RESOLUTION; MANAGEMENT; MODEL;
D O I
10.1142/S0219525909002283
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Providing intelligent algorithms to manage the ever-increasing flow of air traffic is critical to the efficiency and economic viability of air transportation systems. Yet, current automated solutions leave existing human controllers "out of the loop" rendering the potential solutions both technically dangerous (e.g. inability to react to suddenly developing conditions) and politically charged (e.g. role of air traffic controllers in a fully automated system). Instead, this paper outlines a distributed agent-based solution where agents provide suggestions to human controllers. Though conceptually pleasing, this approach introduces two critical research issues. First, the agent actions are now filtered through interactions with other agents, human controllers and the environment before leading to a system state. This indirect action-to-effect process creates a complex learning problem. Second, even in the best case, not all air traffic controllers will be willing or able to follow the agents' suggestions. This partial participation effect will require the system to be robust to the number of controllers that follow the agent suggestions. In this paper, we present an agent reward structure that allows agents to learn good actions in this indirect environment, and explore the ability of those suggestion agents to achieve good system level performance. We present a series of experiments based on real historical air traffic data combined with simulation of air traffic flow around the New York city area. Results show that the agents can improve system-wide performance by up to 20% over that of human controllers alone, and that these results degrade gracefully when the number of human controllers that follow the agents' suggestions declines.
引用
收藏
页码:493 / 512
页数:20
相关论文
共 30 条
[11]  
Bonaceto C., 2005, P 7 INT NDM C AMST N
[12]  
Campbell K. C., 2000, 3 US EUR AIR TRAFF M
[13]  
Christ DD, 2004, BIOTECH PHARM ASPECT, V1, P327
[14]  
Hill J.C., 2005, AAMAS '05, P1083
[15]   On agent-based software engineering [J].
Jennings, NR .
ARTIFICIAL INTELLIGENCE, 2000, 117 (02) :277-296
[16]  
Joint Economic Commission Majority Staff, 2008, YOUR FLIGHT HAS BEEN, P1
[17]  
JONKER G, 2007, P 6 INT JOINT C AUT
[18]  
McNally D., 2006, AIAA GUID NAV CONTR
[19]  
Nagel K, 2002, PEDESTRIAN AND EVACUATION DYNAMICS, P161
[20]   Conflict resolution problems for air traffic management systems solved with mixed integer programming [J].
Pallottino, L ;
Feron, EM ;
Bicchi, A .
IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2002, 3 (01) :3-11