Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems

Conference Paper (2022)
Author(s)

M. Suau (TU Delft - Interactive Intelligence)

J. He (TU Delft - Interactive Intelligence)

Mustafa Mert Çelikok

Matthijs T. J. Spaan (TU Delft - Algorithmics)

F.A. Oliehoek (TU Delft - Interactive Intelligence)

Research Group
Interactive Intelligence
Copyright
© 2022 M. Suau, J. He, Mustafa Mert Çelikok, M.T.J. Spaan, F.A. Oliehoek
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 M. Suau, J. He, Mustafa Mert Çelikok, M.T.J. Spaan, F.A. Oliehoek
Research Group
Interactive Intelligence
ISBN (electronic)
9781713871088
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Due to its high sample complexity, simulation is, as of today, critical for the successful application of reinforcement learning. Many real-world problems, however, exhibit overly complex dynamics, which makes their full-scale simulation computationally slow. In this paper, we show how to factorize large networked systems of many agents into multiple local regions such that we can build separate simulators that run independently and in parallel. To monitor the influence that the different local regions exert on one another, each of these simulators is equipped with a learned model that is periodically trained on real trajectories. Our empirical results reveal that distributing the simulation among different processes not only makes it possible to train large multi-agent systems in just a few hours but also helps mitigate the negative effects of simultaneous learning