Please use this identifier to cite or link to this item:
https://hdl.handle.net/10356/184383
Title: | Automatic landing control for fixed-wing UAV in longitudinal channel based on deep reinforcement learning | Authors: | Li, Jinghang Xu, Shuting Wu, Yu Zhang, Zhe |
Keywords: | Engineering | Issue Date: | 2024 | Source: | Li, J., Xu, S., Wu, Y. & Zhang, Z. (2024). Automatic landing control for fixed-wing UAV in longitudinal channel based on deep reinforcement learning. Drones, 8(10), 568-. https://dx.doi.org/10.3390/drones8100568 | Journal: | Drones | Abstract: | The objective is to address the control problem associated with the landing process of unmanned aerial vehicles (UAVs), with a particular focus on fixed-wing UAVs. The Proportional–Integral–Derivative (PID) controller is a widely used control method, which requires the tuning of its parameters to account for the specific characteristics of the landing environment and the potential for external disturbances. In contrast, neural networks can be modeled to operate under given inputs, allowing for a more precise control strategy. In light of these considerations, a control system based on reinforcement learning is put forth, which is integrated with the conventional PID guidance law to facilitate the autonomous landing of fixed-wing UAVs and the automated tuning of PID parameters through the use of a Deep Q-learning Network (DQN). A traditional PID control system is constructed based on a fixed-wing UAV dynamics model, with the flight state being discretized. The landing problem is transformed into a Markov Decision Process (MDP), and the reward function is designed in accordance with the landing conditions and the UAV’s attitude, respectively. The state vectors are fed into the neural network framework, and the optimized PID parameters are output by the reinforcement learning algorithm. The optimal policy is obtained through the training of the network, which enables the automatic adjustment of parameters and the optimization of the traditional PID control system. Furthermore, the efficacy of the control algorithms in actual scenarios is validated through the simulation of UAV state vector perturbations and ideal gliding curves. The results demonstrate that the controller modified by the DQN network exhibits a markedly superior convergence effect and maneuverability compared to the unmodified traditional controller. | URI: | https://hdl.handle.net/10356/184383 | ISSN: | 2504-446X | DOI: | 10.3390/drones8100568 | Schools: | School of Mechanical and Aerospace Engineering | Rights: | © 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https:// creativecommons.org/licenses/by/ 4.0/). | Fulltext Permission: | open | Fulltext Availability: | With Fulltext |
Appears in Collections: | MAE Journal Articles |
Files in This Item:
File | Description | Size | Format | |
---|---|---|---|---|
drones-08-00568.pdf | 6.85 MB | Adobe PDF | ![]() View/Open |
SCOPUSTM
Citations
50
2
Updated on May 5, 2025
Page view(s)
37
Updated on May 6, 2025
Download(s)
1
Updated on May 6, 2025
Google ScholarTM
Check
Altmetric
Items in DR-NTU are protected by copyright, with all rights reserved, unless otherwise indicated.