Action robust reinforcement learning for air mobility deconfliction against conflict induced spoofing
Date published
Free to read from
Authors
Supervisor/s
Journal Title
Journal ISSN
Volume Title
Publisher
Department
Type
ISSN
Format
Citation
Abstract
Increased dynamic drone usage has increased complexity in aerial navigation and often demands distributed local deconfliction. Due to the high velocities and few landmarks, robust deconfliction relies on precise positioning and synchronization. However, intentional spoofing attacks aimed at inducing navigation conflicts threaten the reliability of conventional techniques. Here, we address these concerns by establishing a baseline on the impact of novel conflict-inducing spoofing attacks on existing geometric navigation methods. Based on the impact of the attacks on the navigation, reinforcement learning (RL) strategy is used to counter the effects of spoofing attacks. In order to counter the effect of spoofing in randomized dynamic airspace conditions, a zero-sum action-robust (ZSAR) RL based on mixed Nash equilibrium objective is used. The proposed methodology yields an improved number of conflict-free paths while reducing average conflicts compared to existing state of the art RL strategies, thus making it suitable for deploying autonomous aircrafts.
Description
Software Description
Software Language
Github
Keywords
DOI
Rights
Relationships
Relationships
Resources
Funder/s
This work was supported in part by the Royal Academy of Engineering and the Office of the Chief Science Adviser for National Security under the UK Intelligence Community Postdoctoral Research under Grant ICRF2324-7-150; in part by the Engineering and Physical Sciences Research Council (EPSRC); and in part by the U.K. Research and Innovation (UKRI), “TAS-S: Trustworthy Autonomous Systems: Security” under Grant EP/V026763/1.