Continuous state and action Q-learning framework applied to quadrotor UAV control

Master Thesis (2017)
Author(s)

A.E. Naruta (TU Delft - Aerospace Engineering)

Contributor(s)

EJ van Kampen – Mentor

Faculty
Aerospace Engineering
Copyright
© 2017 Anton Naruta
More Info
expand_more
Publication Year
2017
Language
English
Copyright
© 2017 Anton Naruta
Graduation Date
08-09-2017
Awarding Institution
Delft University of Technology
Programme
Aerospace Engineering
Faculty
Aerospace Engineering
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This paper describes an implementation of a reinforcement learning-based framework applied to the control of a multi-copter rotorcraft. The controller is based on continuous state and action Q-learning. The policy is stored using a radial basis function neural network. Distance-based neuron activation is used to optimize the generalization algorithm for computational performance. The training proceeds off-line, using a reduced-order model of the controlled system. The model is identified and stored in the form of a neural network. The framework incorporates a dynamics inversion controller, based on the identified model. Simulated flight tests confirm the controller's ability to track the reference state signal and outperform a conventional proportional-derivative(PD) controller. The contributions of the developed framework are a computationally-efficient method to store a $\mathcal{Q}$-function generalization, continuous action selection based on local $\mathcal{Q}$-function approximation and a combination of model identification and offline learning for inner-loop control of a UAV system.

Files

Main.pdf
(pdf | 7.69 Mb)
License info not available