TY - JOUR
T1 - Toward Conflict Resolution with Deep Multi-Agent Reinforcement Learning
AU - Isufaj, Ralvi
AU - Sebastia, David Aranega
AU - Piera, Miquel Angel
N1 - Publisher Copyright:
© 2022 by the American Institute of Aeronautics and Astronautics, Inc. All rights reserved.
PY - 2022
Y1 - 2022
N2 - Safety in air traffic management at the tactical level is ensured by human controllers. Automatic detection and resolution tools are one way to assist controllers in their tasks. However, the majority of existing methods do not account for factors that can affect the quality and efficiency of resolutions. Furthermore, future challenges such as sustainability and the environmental impact of aviation must be tackled. In this work, we propose an innovative approach to pairwise conflict resolution, by modeling it as a multi-agent reinforcement learning to improve the quality of resolutions based on a combination of several factors. We use multi-agent deep deterministic policy gradient to generate resolution maneuvers. We propose a reward function that besides solving the conflicts attempts to optimize the resolutions in terms of time, fuel consumption, and airspace complexity. The models are evaluated on real traffic, with a data augmentation technique utilized to increase the variance of conflict geometries. We achieve promising results with a resolution rate of 93%, without the agents having any previous knowledge of the dynamics of the environment. Furthermore, the agents seem to be able to learn some desirable behaviors such as preferring small heading changes to solve conflicts in one time step. Nevertheless, the nonstationarity of the environment makes the learning procedure nontrivial. We argue ways that tangible qualities such as resolution rate and intangible qualities such as resolution acceptability and explainability can be improved.
AB - Safety in air traffic management at the tactical level is ensured by human controllers. Automatic detection and resolution tools are one way to assist controllers in their tasks. However, the majority of existing methods do not account for factors that can affect the quality and efficiency of resolutions. Furthermore, future challenges such as sustainability and the environmental impact of aviation must be tackled. In this work, we propose an innovative approach to pairwise conflict resolution, by modeling it as a multi-agent reinforcement learning to improve the quality of resolutions based on a combination of several factors. We use multi-agent deep deterministic policy gradient to generate resolution maneuvers. We propose a reward function that besides solving the conflicts attempts to optimize the resolutions in terms of time, fuel consumption, and airspace complexity. The models are evaluated on real traffic, with a data augmentation technique utilized to increase the variance of conflict geometries. We achieve promising results with a resolution rate of 93%, without the agents having any previous knowledge of the dynamics of the environment. Furthermore, the agents seem to be able to learn some desirable behaviors such as preferring small heading changes to solve conflicts in one time step. Nevertheless, the nonstationarity of the environment makes the learning procedure nontrivial. We argue ways that tangible qualities such as resolution rate and intangible qualities such as resolution acceptability and explainability can be improved.
UR - http://www.scopus.com/inward/record.url?scp=85133723275&partnerID=8YFLogxK
U2 - 10.2514/1.D0296
DO - 10.2514/1.D0296
M3 - Article
AN - SCOPUS:85133723275
SN - 2380-9450
VL - 30
SP - 71
EP - 80
JO - Journal of Air Transportation
JF - Journal of Air Transportation
IS - 3
ER -