ARMCHAIR

Integrated Inverse Reinforcement Learning and Model Predictive Control for Human-Robot Collaboration

Journal Article (2025)
Author(s)

Angelo Caregnato-Neto (Instituto Tecnológico de Aeronáutica)

L. Cavalcante Siebert (TU Delft - Interactive Intelligence)

A. Zgonnikov (TU Delft - Human-Robot Interaction)

Marcos R.O.A. Maximo (Instituto Tecnológico de Aeronáutica)

Rubens J.M. Afonso (Instituto Tecnológico de Aeronáutica)

Research Group
Interactive Intelligence
DOI related publication
https://doi.org/10.1007/s12369-025-01332-4
More Info
expand_more
Publication Year
2025
Language
English
Research Group
Interactive Intelligence
Bibliographical Note
Green Open Access added to TU Delft Institutional Repository 'You share, we take care!' - Taverne project https://www.openaccess.nl/en/publishing/publisher-deals Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public.@en
Issue number
12
Volume number
17
Pages (from-to)
3173-3189
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

One of the key issues in human-robot collaboration is the development of computational models that allow robots to predict and adapt to human behavior. Much progress has been achieved in developing such models, as well as control techniques that address the autonomy problems of motion planning and decision-making in robotics. However, the integration of computational models of human behavior with such control techniques still poses a major challenge, resulting in a bottleneck for efficient collaborative human-robot teams. In this context, we present a novel architecture for human-robot collaboration: Adaptive Robot Motion for Collaboration with Humans using Adversarial Inverse Reinforcement learning (ARMCHAIR). Our solution leverages adversarial inverse reinforcement learning and model predictive control to compute optimal trajectories and decisions for a mobile multi-robot system that collaborates with a human in an exploration task. During the mission, ARMCHAIR operates without human intervention, autonomously identifying the necessity to support and acting accordingly. Our approach also explicitly addresses the network connectivity requirement of the human-robot team. Extensive simulation-based evaluations demonstrate that ARMCHAIR allows a group of robots to safely support a simulated human in an exploration scenario, preventing collisions and network disconnections, and improving the overall performance of the task.

Files

S12369-025-01332-4.pdf
(pdf | 4.28 Mb)
License info not available
warning

File under embargo until 20-04-2026