UAV Control Method Combining Reptile Meta-Reinforcement Learning and Generative Adversarial Imitation Learning
Article
Article Title | UAV Control Method Combining Reptile Meta-Reinforcement Learning and Generative Adversarial Imitation Learning |
---|---|
ERA Journal ID | 212586 |
Article Category | Article |
Authors | Jiang, Shui, Ge, Yanning, Yang, Xu, Yang, Wencheng and Cui, Hui |
Journal Title | Future Internet |
Journal Citation | 16 (3) |
Number of Pages | 18 |
Year | 2024 |
Publisher | MDPI AG |
Place of Publication | Switzerland |
ISSN | 1999-5903 |
Digital Object Identifier (DOI) | https://doi.org/10.3390/fi16030105 |
Web Address (URL) | https://www.mdpi.com/1999-5903/16/3/105 |
Abstract | Reinforcement learning (RL) is pivotal in empowering Unmanned Aerial Vehicles (UAVs) to navigate and make decisions efficiently and intelligently within complex and dynamic surroundings. Despite its significance, RL is hampered by inherent limitations such as low sample efficiency, restricted generalization capabilities, and a heavy reliance on the intricacies of reward function design. These challenges often render single-method RL approaches inadequate, particularly in the context of UAV operations where high costs and safety risks in real-world applications cannot be overlooked. To address these issues, this paper introduces a novel RL framework that synergistically integrates meta-learning and imitation learning. By leveraging the Reptile algorithm from meta-learning and Generative Adversarial Imitation Learning (GAIL), coupled with state normalization techniques for processing state data, this framework significantly enhances the model’s adaptability. It achieves this by identifying and leveraging commonalities across various tasks, allowing for swift adaptation to new challenges without the need for complex reward function designs. To ascertain the efficacy of this integrated approach, we conducted simulation experiments within both two-dimensional environments. The empirical results clearly indicate that our GAIL-enhanced Reptile method surpasses conventional single-method RL algorithms in terms of training efficiency. This evidence underscores the potential of combining meta-learning and imitation learning to surmount the traditional barriers faced by reinforcement learning in UAV trajectory planning and decision-making processes. |
Keywords | unmanned aerial vehicles (UAVs); meta-reinforcement learning; enerative adversarial imitation learning |
Contains Sensitive Content | Does not contain sensitive content |
ANZSRC Field of Research 2020 | 460304. Computer vision |
Byline Affiliations | Fujian Normal University, China |
Minjiang University, China | |
School of Mathematics, Physics and Computing | |
Monash University |
https://research.usq.edu.au/item/z5y70/uav-control-method-combining-reptile-meta-reinforcement-learning-and-generative-adversarial-imitation-learning
Download files
45
total views30
total downloads1
views this month2
downloads this month