Reinforcement Learning Methods for Fixed-Wing Aircraft Control | IEEE Conference Publication | IEEE Xplore

Reinforcement Learning Methods for Fixed-Wing Aircraft Control


Abstract:

In recent years, reinforcement learning (RL) has been widely researched for fixed-wing aircraft control, offering transformative potential for more adaptive, efficient, a...Show More

Abstract:

In recent years, reinforcement learning (RL) has been widely researched for fixed-wing aircraft control, offering transformative potential for more adaptive, efficient, and autonomous flight operations. This paper presents a comprehensive analysis of the current state, advantages, challenges, and future prospects of employing RL in fixed-wing aircraft control. The advantages of RL include model-free, adaptive decision-making, and effectively solving complex nonlinear constraints, leading to increased efficiency and maneuverability in high-dynamic environments. However, significant challenges, such as the guarantee of stability and optimization, interpretability, and sim-to-real transferring, highlight areas necessitating further attention and development. Overcoming these challenges requires collaborative efforts among aviation experts, control theory experts, and machine learning researchers to ensure safe, reliable integration of RL in flight systems. As research and advancements continue, the RL holds promise for providing a more efficient, maneuverable, and autonomous strategy for controlling fixed-wing aircraft in the future.
Date of Conference: 16-18 December 2023
Date Added to IEEE Xplore: 09 February 2024
ISBN Information:
Conference Location: Changsha, China

Funding Agency:

References is not available for this document.

I. Introduction

Reinforcement learning [39] has emerged as an effective approach in the development of autonomous systems, offering promising avenues for enhancing control mechanisms in various domains, such as autonomous driving [41], robotic manipulation [19], and quadrotor flight control [21]. In addition, the integration of RL in the control of the fixed-wing aircraft represents a significant stride toward adaptive and autonomous flight operations. Historically, fixed-wing aircraft control systems have heavily relied on pre-programmed algorithms, remote control, and traditional control methods (e.g. Proportional-Integral-Derivative (PID) control) [2]. However, the landscape is rapidly evolving with the advent of artificial intelligence and machine learning. As a subset of machine learning, RL offers a distinct advantage in this sphere by enabling systems to learn from their experiences and interactions within dynamic environments. Autonomous flight control demands continuous decision-making in response to real-time changes in flight, such as task reallocation and wind disturbance. The application of RL in this context aims to imbue aircraft control systems with the ability to adapt, learn, and optimize their decision-making processes from the feedback of the environment. The fundamental principle of RL involves the interaction of an agent, in this case, the aircraft control system, with an environment. Through this interaction, the system learns to perform sequences of actions that result in maximizing a cumulative reward. In the context of fixed-wing aircraft control, this learning process could involve attitude control [3], autopilot control [11], managing control surfaces [27], or even responding to unexpected situations without direct human intervention [34]. The potential benefits of applying RL in aircraft control are manifold. The adaptive nature of RL allows systems to learn and adjust their behavior in response to novel situations or changes in the environment. For example, the environment in which the aircraft operates is often uncertain, such as changes in air density, wind direction, and intensity. Additionally, the parameters of aircraft systems may change over time or be affected by external factors. For another example, in a tense flight environment, such as pursuit and escape, flight instructions are often switched frequently. This adaptability is particularly crucial to making swift and accurate decision-making to address the above unforeseen conditions. However, the integration of RL in aircraft control is challenging. Safety, reliability, interpretability of models, and adherence to stringent aviation regulations pose significant hurdles. Addressing these challenges is imperative for the successful adoption of RL in ensuring the safety and efficiency of the fixed-wing aircraft. In light of these challenges and the remarkable potential, the implementation of RL in controlling fixed-wing aircraft has infinite possibilities. The ongoing scientific research in this field hold the promise of transforming how aircraft control systems operate, leading to safer, more adaptive, and more efficient flight operations.

Select All
1.
J H Bae, H Jung, S Kim, S Kim and Y D. Kim, "Deep Reinforcement Learning-Based Air-to-Air Combat Maneuver Generation in a Realistic Environment", IEEE Access, vol. 11, pp. 26427-26440, 2023.
2.
Randal W Beard and Timothy W McLain, Small unmanned aircraft: Theory and practice, Princeton university press, 2012.
3.
E Bohn, E M Coates, S Moe and T A Johansen, "Deep Reinforcement Learning Attitude Control of Fixed-Wing UAVs Using Proximal Policy optimization", 2019 International Conference on Unmanned Aircraft Systems (ICUAS), pp. 523-533, 2019.
4.
Eivind Bohn, Erlend M Coates, Dirk Reinhardt and Tor Arne Johansen, "Data-Efficient Deep Reinforcement Learning for Attitude Control of Fixed-Wing UAVs: Field Experiments", IEEE Transactions on Neural Networks and Learning Systems, 2023.
5.
Ralph Allan Bradley and Milton E Terry, "Rank analysis of incomplete block designs: I. The method of paired comparisons", Biometrika, vol. 39, no. 3/4, pp. 324-345, 1952.
6.
J Chai, W Chen, Y Zhu, Z X. Yao and D Zhao, "A Hierarchical Deep Reinforcement Learning Framework for 6-DOF UCAV Air-to-Air Combat", IEEE Transactions on Systems Man and Cybernetics: Systems, pp. 1-13, 2023.
7.
Jongkwan Choi, Hyeon Min Kim, Ha Jun Hwang, Yong-Duk Kim and Chang Ouk Kim, "Modular Reinforcement Learning for Autonomous UAV Flight Control", Drones, vol. 7, no. 7, pp. 418, 2023.
8.
Robert J Clarke, Liam Fletcher, Colin Greatwood, Antony Waldock and Thomas S Richardson, "Closed-loop Q-learning control of a small unmanned aircraft", AIAA Scitech 2020 Forum, pp. 1234, 2020.
9.
Shanelle G Clarke and Inseok Hwang, "Deep reinforcement learning control for aerobatic maneuvering of agile fixed-wing aircraft", AIAA Scitech 2020 Forum, pp. 136, 2020.
10.
Francesco D’Apolito, "Reinforcement Learning Training Environment for Fixed Wing UAV Collision Avoidance", IFAC-PapersOnLine, vol. 55, no. 39, pp. 281-285, 2022.
11.
Agostino De Marco, Paolo Maria D’Onza and Sabato Manfredi, "A deep reinforcement learning control approach for high-performance aircraft", Nonlinear Dynamics, vol. 111, no. 18, pp. 17037-17077, 2023.
12.
Raphael C Engelhardt, Marc Oedingen, Moritz Lange, Laurenz Wiskott and Wolfgang Konen, "Iterative Oblique Decision Trees Deliver Explainable RL Models", Algorithms, vol. 16, no. 6, pp. 282, 2023.
13.
Liam J Fletcher, Robert J Clarke, Thomas S Richardson and Mark Hansen, "Reinforcement learning for a perched landing in the presence of wind", AIAA Scitech 2021 Forum, pp. 1282, 2021.
14.
Bishwamittra Ghosh and Kuldeep S Meel, "IMLI: An incremental framework for MaxSAT-based learning of interpretable classification rules", Proceedings of the 2019 AAAI/ACM Conference on AI Ethics and Society, pp. 203-210, 2019.
15.
Tuomas Haarnoja, Aurick Zhou, Pieter Abbeel and Sergey Levine, "Soft actor-critic: Off-policy maximum entropy deep reinforcement learning with a stochastic actor", International Conference on Machine Learning, pp. 1861-1870, 2018.
16.
Mark Hammond, "Deep reinforcement learning in the enterprise: Bridging the gap from games to industry", Artificial Intelligence Conference Presentation, 2017.
17.
Minghao Han, Lixian Zhang, Jun Wang and Wei Pan, "Actor-critic reinforcement learning for control with stability guarantee", IEEE Robotics and Automation Letters, vol. 5, no. 4, pp. 6217-6224, 2020.
18.
Weijun Hu, Zhiqiang Gao, Jiale Quan, Xianlong Ma, Jingyi Xiong and Weijie Zhang, "Fixed-wing stalled maneuver control technology based on deep reinforcement learning", 2022 IEEE 5th International Conference on Big Data and Artificial Intelligence (BDAI), pp. 19-25, 2022.
19.
Boris Ivanovic, James Harrison, Apoorva Sharma, Mo Chen and Marco Pavone, "Barc: Backward reachability curriculum for robotic reinforcement learning", 2019 International Conference on Robotics and Automation (ICRA), pp. 15-21, 2019.
20.
Jithin Jagannath, Anu Jagannath, Sean Furman and Tyler Gwin, "Deep learning and reinforcement learning for autonomous unmanned aerial systems: Roadmap for theory to deployment", Deep Learning for Unmanned Systems, pp. 25-82, 2021.
21.
Elia Kaufmann, Leonard Bauersfeld, Antonio Loquercio, Matthias Müller, Vladlen Koltun and Davide Scaramuzza, "Champion-level drone racing using deep reinforcement learning", Nature, vol. 620, no. 7976, pp. 982-987, 2023.
22.
R Kevin, "Regret-based reward elicitation for Markov decision processes", PhD thesis, 2014.
23.
Stephen Kimathi, "Application of reinforcement learning in heading control of a fixed wing uav using x-plane platform", 2017.
24.
Ladislav Kopečný, Jakub Hnidka and Josef Bajer, "Preliminary design of a fixed-winged UAV autopilot for take-off phase and its assessment", 2022 20th International Conference on Mechatronics-Mechatronika (ME), pp. 1-6, 2022.
25.
Shashank Kumar, "A brief review on Unmanned Combat Aerial Vehicle (UCAV)", Available at SSRN 3593220, 2020.
26.
L Li, Z Zhou, J Chai, Z Liu, Y Zhu and J Yi, "Learning Continuous 3-DoF Air-to-Air Close-in Combat Strategy using Proximal Policy Optimization", 2022 IEEE Conference on Games (CoG), pp. 616-619, 2022.
27.
Lun Li, Xuebo Zhang, Chenxu Qian and Runhua Wang, "Basic flight maneuver generation of fixed-wing plane based on proximal policy optimization", Neural Computing and Applications, 2023.
28.
Timothy P Lillicrap, Jonathan J Hunt, Alexander Pritzel, Nicolas Heess, Tom Erez, Yuval Tassa, et al., "Continuous control with deep reinforcement learning", 2015.
29.
Runze Liu, Fengshuo Bai, Yali Du and Yaodong Yang, "Meta-reward-net: Implicitly differentiable reward learning for preference-based reinforcement learning", Advances in Neural Information Processing Systems, vol. 35, pp. 22270-22284, 2022.
30.
Xiaoteng Ma, Li Xia and Qianchuan Zhao, "Air-combat strategy using deep Q-learning", 2018 Chinese Automation Congress (CAC), pp. 3952-3957, 2018.
Contact IEEE to Subscribe

References

References is not available for this document.