Vision-Based Reinforcement Learning for the guidance of an AR Drone 2

Master Thesis (2018)
Author(s)

M. Siddiquee (TU Delft - Aerospace Engineering)

Contributor(s)

E. van Kampen – Mentor

Jaime Junell – Mentor

Faculty
Aerospace Engineering
Copyright
© 2018 Manan Siddiquee
More Info
expand_more
Publication Year
2018
Language
English
Copyright
© 2018 Manan Siddiquee
Graduation Date
17-04-2018
Awarding Institution
Delft University of Technology
Programme
Aerospace Engineering
Related content

Wikipedia article describing the use of PaparazziUAV and FlightGear for vision-based simulation as used in this study.

https://wiki.paparazziuav.org/wiki/Load_Screenshots_from_FlightGear

Video demonstrating the learned behavior of the agent

https://www.youtube.com/watch?v=Q_soXNEBaZI
Faculty
Aerospace Engineering
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Reinforcement Learning (RL) has been applied to teach quadcopters guidance tasks. Most applications rely on position information from an absolute reference
system such as Global Positioning System (GPS). The dependence on "absolute
position" information is a general limitation in the autonomous flight of Unmanned Aerial Vehicles (UAVs). Environments that have weak or no GPS signals are difficult to traverse for them. Instead of using absolute position, it is
possible to sense the environment and the information contained within it in
order to come up with a "relative" description of the UAV's position. This
paper presents the design of a RL agent with relative vision-based states and rewards for the teaching of a guidance task to a quadcopter. The agent is taught the task of turning towards a red marker and approaching it in simulation and in flight tests. A more complex task of travelling between a blue and a red marker is trained in simulation. This work shows that relative vision-based states and rewards can be used with RL to teach quadcopters simple guidance tasks. The performance of the trained agent is inconsistent in simulation and flight test due to the inherent partial
observability in the relative description of the state.

Files

License info not available