Using Deep Reinforcement Learning to Improve the Robustness of UAV Lateral-Directional Control

Conference Paper (2022)
Author(s)

Rui Wang (Northwestern Polytechnical University)

Zhou Zhou (Northwestern Polytechnical University)

Xiaoping Zhu (Northwestern Polytechnical University)

L. Zheng (TU Delft - Control & Simulation)

Research Group
Control & Simulation
More Info
expand_more
Publication Year
2022
Language
English
Research Group
Control & Simulation
Pages (from-to)
5489-5504
ISBN (electronic)
9781713871163
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

For a small low-cost Unmanned Aerial Vehicle (UAV), the accurate aerodynamics and flight dynamics characteristics wouldn't be obtained easily, and the control coupling is serious, so the robustness of its flight controller must be considered carefully. In order to solve the problem, a Lateral-Directional (Lat-Dir) flight control method based on Deep Reinforcement Learning (DRL) are proposed in this paper. Firstly, based on the nominal state, three control laws are designed: classical Proportional Integral Derivative (PID) control, Linear Quadratic Gaussian (LQG) control based on modern control theory, and Deep Reinforcement Learning (DRL) control based on Twin Delayed Deep Deterministic Policy Gradient (TD3) method. In order to solve the problem of incomprehensible physical meaning of neural network in DRL, a simplified control strategy network is derived based on the inspiration of PID controller. In order to solve the problem that the reward function of DRL is difficult to be determined, the weights of the optimal quadratic function designed by LQG method are adopted, and the weights of control output considering discretization is added also. Then, the three controller are applied to nominal flight state and deviation state respectively, and the numerical flight simulation is carried out. The results show that, in the nominal state, the performance of DRL is close to the LQG and better than the PID. In the deviation state, which the lateral and directional static stable derivatives are changed artificially from stable to neutral stable, the rise time and adjustment time of the DRL change slightly, while the LQG degrades seriously and appears instable, and it is proved that the proposed DRL control method has better performance robustness.

Files

ICAS2022_0065_paper.pdf
(pdf | 1.8 Mb)
License info not available