Toward Scalable Multirobot Control

Fast Policy Learning in Distributed MPC

Journal Article (2025)
Author(s)

Xinglong Zhang (National University of Defense Technology)

W. Pan (TU Delft - Robot Dynamics)

Cong Li (National University of Defense Technology)

Xin Xu (National University of Defense Technology)

Xiangke Wang (National University of Defense Technology)

Ronghua Zhang (National University of Defense Technology)

Dewen Hu (National University of Defense Technology)

Research Group
Robot Dynamics
DOI related publication
https://doi.org/10.1109/TRO.2025.3531818
More Info
expand_more
Publication Year
2025
Language
English
Research Group
Robot Dynamics
Bibliographical Note
Green Open Access added to TU Delft Institutional Repository 'You share, we take care!' - Taverne project https://www.openaccess.nl/en/you-share-we-take-care Otherwise as indicated in the copyright section: the publisher is the copyright holder of this work and the author uses the Dutch legislation to make this work public.@en
Volume number
41
Pages (from-to)
1491-1512
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Distributed model predictive control (DMPC) is promising in achieving optimal cooperative control in multirobot systems (MRS). However, real-time DMPC implementation relies on numerical optimization tools to periodically calculate local control sequences online. This process is computationally demanding and lacks scalability for large-scale, nonlinear MRS. This article proposes a novel distributed learning-based predictive control framework for scalable multirobot control. Unlike conventional DMPC methods that calculate open-loop control sequences, our approach centers around a computationally fast and efficient distributed policy learning algorithm that generates explicit closed-loop DMPC policies for MRS without using numerical solvers. The policy learning is executed incrementally and forward in time in each prediction interval through an online distributed actor-critic implementation. The control policies are successively updated in a receding-horizon manner, enabling fast and efficient policy learning with the closed-loop stability guarantee. The learned control policies could be deployed online to MRS with varying robot scales, enhancing scalability and transferability for large-scale MRS. Furthermore, we extend our methodology to address the multirobot safe learning challenge through a force field-inspired policy learning approach. We validate our approach's effectiveness, scalability, and efficiency through extensive experiments on cooperative tasks of large-scale wheeled robots and multirotor drones. Our results demonstrate the rapid learning and deployment of DMPC policies for MRS with scales up to 10 000 units.

Files

Toward_Scalable_Multirobot_Con... (pdf)
(pdf | 5.36 Mb)
- Embargo expired in 20-06-2025
License info not available