ILeSiA

Interactive Learning of Robot Situational Awareness From Camera Input

Journal Article (2025)
Author(s)

P. Vanc (Czech Technical University, TU Delft - Learning & Autonomous Control)

G. Franzese (TU Delft - Learning & Autonomous Control)

Jan K. Behrens (Czech Technical University)

C. Della Santina (TU Delft - Learning & Autonomous Control)

Karla Stepanova (Czech Technical University)

J. Kober (TU Delft - Learning & Autonomous Control)

R. Babuska (Czech Technical University, TU Delft - Learning & Autonomous Control)

Research Group
Learning & Autonomous Control
DOI related publication
https://doi.org/10.1109/LRA.2025.3601037
More Info
expand_more
Publication Year
2025
Language
English
Research Group
Learning & Autonomous Control
Issue number
10
Volume number
10
Pages (from-to)
10490-10497
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Learning from demonstration is a promising approach for teaching robots new skills. However, a central challenge in the execution of acquired skills is the ability to recognize faults and prevent failures. This is essential because demonstrations typically cover only a limited set of scenarios and often only the successful ones. During task execution, unforeseen situations may arise, such as changes in the robot's environment or interaction with human operators. To recognize such situations, this paper focuses on teaching the robot situational awareness by using a camera input and labeling frames as safe or risky. We train a Gaussian Process (GP) regression model fed by a low-dimensional latent space representation of the input images. The model outputs a continuous risk score ranging from zero to one, quantifying the degree of risk at each timestep. This allows for pausing task execution in unsafe situations and directly adding new training data, labeled by the human user. Our experiments on a robotic manipulator show that the proposed method can reliably detect both known and novel faults using only a single example for each new fault. In contrast, a standard multi-layer perceptron (MLP) performs well only on faults it has encountered during training. Our method enables the next generation of cobots to be rapidly deployed with easy-to-set-up, vision-based risk assessment, proactively safeguarding humans and detecting misaligned parts or missing objects before failures occur.