Speeding up Deep Reinforcement Learning through Influence-Augmented Local Simulators

Conference Paper (2022)
Author(s)

Miguel Suau (TU Delft - Interactive Intelligence)

Jinke He (TU Delft - Interactive Intelligence)

Matthijs T. J. Spaan (TU Delft - Algorithmics)

Frans Oliehoek (TU Delft - Interactive Intelligence)

Research Group
Interactive Intelligence
Copyright
© 2022 M. Suau, J. He, M.T.J. Spaan, F.A. Oliehoek
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 M. Suau, J. He, M.T.J. Spaan, F.A. Oliehoek
Research Group
Interactive Intelligence
Pages (from-to)
1735-1737
ISBN (electronic)
978-171385433-3
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Learning effective policies for real-world problems is still an open challenge for the field of reinforcement learning (RL). The main limitation being the amount of data needed and the pace at which that data can be obtained. In this paper, we study how to build lightweight simulators of complicated systems that can run sufficiently fast for deep RL to be applicable. We focus on domains where agents interact with a reduced portion of a larger environment while still being affected by the global dynamics. Our method combines the use of local simulators with learned models that mimic the influence of the global system. The experiments reveal that incorporating this idea into the deep RL workflow can considerably accelerate the training process and presents several opportunities for the future.

Files

3535850.3536093.pdf
(pdf | 1.44 Mb)
- Embargo expired in 05-12-2022
License info not available