Visual Navigation in Real-World Indoor Environments Using End-to-End Deep Reinforcement Learning

Journal Article (2021)
Author(s)

Jonas Kulhanek (Czech Technical University)

Erik Derner (Czech Technical University)

R Babuska (TU Delft - Learning & Autonomous Control, Czech Technical University)

Research Group
Learning & Autonomous Control
Copyright
© 2021 Jonas Kulhanek, Erik Derner, R. Babuska
DOI related publication
https://doi.org/10.1109/LRA.2021.3068106
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 Jonas Kulhanek, Erik Derner, R. Babuska
Research Group
Learning & Autonomous Control
Issue number
3
Volume number
6
Pages (from-to)
4345-4352
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Visual navigation is essential for many applications in robotics, from manipulation, through mobile robotics to automated driving. Deep reinforcement learning (DRL) provides an elegant map-free approach integrating image processing, localization, and planning in one module, which can be trained and therefore optimized for a given environment. However, to date, DRL-based visual navigation was validated exclusively in simulation, where the simulator provides information that is not available in the real world, e.g., the robot's position or segmentation masks. This precludes the use of the learned policy on a real robot. Therefore, we present a novel approach that enables a direct deployment of the trained policy on real robots. We have designed a new powerful simulator capable of domain randomization. To facilitate the training, we propose visual auxiliary tasks and a tailored reward scheme. The policy is fine-tuned on images collected from real-world environments. We have evaluated the method on a mobile robot in a real office environment. The training took approximately 30 hours on a single GPU. In 30 navigation experiments, the robot reached a 0.3-meter neighbourhood of the goal in more than 86.7% of cases. This result makes the proposed method directly applicable to tasks like mobile manipulation.

Files

09384194.pdf
(pdf | 4.24 Mb)
License info not available