Uncovering reward goals in distributed drone swarms using physics-informed multiagent inverse reinforcement learning
dc.contributor.author | Perrusquía, Adolfo | |
dc.contributor.author | Guo, Weisi | |
dc.date.accessioned | 2025-02-27T11:10:58Z | |
dc.date.available | 2025-02-27T11:10:58Z | |
dc.date.freetoread | 2025-02-27 | |
dc.date.issued | 2025-01-01 | |
dc.date.pubOnline | 2024-11-13 | |
dc.description.abstract | The cooperative nature of drone swarms poses risks in the smooth operation of services and the security of national facilities. The control objective of the swarm is, in most cases, occluded due to the complex behaviors observed in each drone. It is paramount to understand which is the control objective of the swarm, whilst understanding better how they communicate with each other to achieve the desired task. To solve these issues, this article proposes a physics-informed multiagent inverse reinforcement learning (PI-MAIRL) that: 1) infers the control objective function or reward function from observational data and 2) uncover the network topology by exploiting a physics-informed model of the dynamics of each drone. The combined contribution enables to understand better the behavior of the swarm, whilst enabling the inference of its objective for experience inference and imitation learning. A physically uncoupled swarm scenario is considered in this study. The incorporation of the physics-informed element allows to obtain an algorithm that is computationally more efficient than model-free IRL algorithms. Convergence of the proposed approach is verified using Lyapunov recursions on a global Riccati equation. Simulation studies are carried out to show the benefits and challenges of the approach. | |
dc.description.journalName | IEEE Transactions on Cybernetics | |
dc.format.extent | pp. 14-23 | |
dc.identifier.citation | Perrusquía A, Guo W. (2025) Uncovering reward goals in distributed drone swarms using physics-informed multiagent inverse reinforcement learning. IEEE Transactions on Cybernetics, Volume 55, Issue 1, January 2025, pp. 14-23 | |
dc.identifier.eissn | 2168-2275 | |
dc.identifier.elementsID | 559341 | |
dc.identifier.issn | 2168-2267 | |
dc.identifier.issueNo | 1 | |
dc.identifier.uri | https://doi.org/10.1109/tcyb.2024.3489967 | |
dc.identifier.uri | https://dspace.lib.cranfield.ac.uk/handle/1826/23525 | |
dc.identifier.volumeNo | 55 | |
dc.language.iso | en | |
dc.publisher | IEEE | |
dc.publisher.uri | https://ieeexplore.ieee.org/document/10752585 | |
dc.rights | Attribution 4.0 International | en |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | |
dc.subject | Drone swarms | |
dc.subject | imitation learning | |
dc.subject | multiagent inverse reinforcement learning (IRL) | |
dc.subject | network topology | |
dc.subject | physics-informed | |
dc.subject | reward function | |
dc.subject | 46 Information and Computing Sciences | |
dc.subject | 4602 Artificial Intelligence | |
dc.subject | Behavioral and Social Science | |
dc.subject | Basic Behavioral and Social Science | |
dc.title | Uncovering reward goals in distributed drone swarms using physics-informed multiagent inverse reinforcement learning | |
dc.type | Article | |
dc.type.subtype | Journal Article | |
dcterms.dateAccepted | 2024-10-30 |