Continuous state and action Q-learning framework applied to quadrotor UAV control

Conference Paper (2019)
Author(s)

Anton Naruta (Student TU Delft)

T Mannucci (TU Delft - Control & Simulation)

Erik-Jan van Kampen (TU Delft - Control & Simulation)

Research Group
Control & Simulation
Copyright
© 2019 Anton Naruta, T. Mannucci, E. van Kampen
DOI related publication
https://doi.org/10.2514/6.2019-0145
More Info
expand_more
Publication Year
2019
Language
English
Copyright
© 2019 Anton Naruta, T. Mannucci, E. van Kampen
Research Group
Control & Simulation
ISBN (electronic)
978-1-62410-578-4
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This paper describes an implementation of a reinforcement learning-based framework applied to the control of a multi-copter rotorcraft. The controller is based on continuous state and action Q-learning. The policy is stored using a radial basis function neural network. Distance-based neuron activation is used to optimize the generalization algorithm for computational performance. The training proceeds off-line, using a reduced-order model of the controlled system. The model is identified and stored in the form of a neural network. The framework incorporates a dynamics inversion controller, based on the identified model. Simulated flight tests confirm the controller’s ability to track the reference state signal and outperform a conventional proportional-derivative(PD) controller. The contributions of the developed framework are a computationally-efficient method to store a Q-function generalization, continuous action selection based on local Q-function approximation and a combination of model identification and offline learning for inner-loop control of a UAV system.

Files

AIAA_2019_Anton_Naruta.pdf
(pdf | 2.64 Mb)
License info not available