Imitation Learning with Inconsistent Demonstrations through Uncertainty-based Data Manipulation

Conference Paper (2021)
Author(s)

Peter Valletta (Student TU Delft)

Rodrigo Perez Dattari (TU Delft - Learning & Autonomous Control)

J. Kober (TU Delft - Learning & Autonomous Control)

Research Group
Learning & Autonomous Control
Copyright
© 2021 Peter Valletta, R.J. Perez Dattari, J. Kober
DOI related publication
https://doi.org/10.1109/ICRA48506.2021.9561686
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 Peter Valletta, R.J. Perez Dattari, J. Kober
Research Group
Learning & Autonomous Control
Pages (from-to)
3655-3661
ISBN (electronic)
978-1-7281-9077-8
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Aleatoric uncertainty estimation, based on the observed training data, is applied for the detection of conflicts in a demonstration data set. The particular focus of this paper is the resolution of conflicting data resulting from scenarios with equivalent action choices, such as obstacle avoidance, path planning or multiple joint configurations. In terms of the estimated uncertainty, the proposed algorithm aims to decrease this otherwise irreducible value through direct alteration of the accrued data set and to provide data that a policy-learning neural network is able to fit appropriately. The proposed algorithm was validated with real robot scenarios while learning from inconsistent demonstrations, where the resulting policies consistently achieved their prescribed objectives. A video showing our method and experiments can be found at: https://youtu.be/oGYnzlW9Ncw.

Files

Imitation_Learning_with_Incons... (pdf)
(pdf | 3.96 Mb)
- Embargo expired in 18-04-2022
License info not available