Print Email Facebook Twitter Improving DRL Of Vision-Based Navigation By Stereo Image Prediction Title Improving DRL Of Vision-Based Navigation By Stereo Image Prediction Author den Ridder, Luc (TU Delft Aerospace Engineering) Contributor de Croon, G.C.H.E. (mentor) Wu, Y. (mentor) Degree granting institution Delft University of Technology Corporate name Delft University of Technology Programme Aerospace Engineering Date 2023-06-20 Abstract Although deep reinforcement learning (DRL) is a highly promising approach to learning robotic vision-based control, it is plagued by long training times. This report introduces a DRL setup that relies on self-supervised learning for extracting depth information valuable for navigation. Specifically, a literature study is conducted to investigate the effects of learning how to synthesize one view from the other in a stereo-vision setup without relying on any preliminary knowledge of the camera extrinisics and how it can be integrated for its downstream use for an obstacle avoidance task. As such, the literature study concludes that competitive geometry-free monocular-to-stereo image view synthesis is feasible due to recent developments in computer vision. The scientific paper further develops concepts proposed in the literature study and benchmarks the proposed architectures on depth estimation benchmarks for KITTI. Competitive results are achieved for view synthesis and despite sub-optimal performance compared to state-of-the-art monocular depth estimation, an ability to encode depth and detect shapes is present and, therefore, satisfactory for the application to DRL. Additionally, the research examines the benefits of using the latent space of a view synthesis architecture compared to other feature extractor methods as an input to the PPO agent implemented as auxiliary tasks. This method achieves quicker convergence and better performance for an obstacle avoidance task in a simulated indoor environment than the autoencoding feature extractor and end-to-end DRL methods. It is only outperformed by the monocular depth estimation feature extractor method. Overall, this research provides valuable insights for developing more efficient and effective DRL methods for monocular camera-based drones. Finally, the complementary code for this research can be found: \url{https://github.com/ldenridder/drl-obstacle-avoidance-view-synthesis}. Subject Autonomous NavigationUAVDeep Reinforcement LearningSelf-supervised learningAuxiliary tasksMonocular VisionDepth EstimationFeature Extractionsimulation To reference this document use: http://resolver.tudelft.nl/uuid:ef354713-924e-4907-a44f-95b67efa638e Part of collection Student theses Document type master thesis Rights © 2023 Luc den Ridder Files PDF MSc_Thesis_Final_Report_L ... Ridder.pdf 28.93 MB Close viewer /islandora/object/uuid:ef354713-924e-4907-a44f-95b67efa638e/datastream/OBJ/view