Analysing factorizations of action-value networks for cooperative multi-agent reinforcement learning

Journal Article (2021)
Author(s)

Jacopo Castellini (University of Liverpool)

Frans Oliehoek (TU Delft - Interactive Intelligence)

Rahul Savani (University of Liverpool)

Shimon Whiteson (University of Oxford)

Research Group
Interactive Intelligence
Copyright
© 2021 Jacopo Castellini, F.A. Oliehoek, Rahul Savani, Shimon Whiteson
DOI related publication
https://doi.org/10.1007/s10458-021-09506-w
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 Jacopo Castellini, F.A. Oliehoek, Rahul Savani, Shimon Whiteson
Research Group
Interactive Intelligence
Issue number
2
Volume number
35
Pages (from-to)
1-53
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Recent years have seen the application of deep reinforcement learning techniques to cooperative multi-agent systems, with great empirical success. However, given the lack of theoretical insight, it remains unclear what the employed neural networks are learning, or how we should enhance their learning power to address the problems on which they fail. In this work, we empirically investigate the learning power of various network architectures on a series of one-shot games. Despite their simplicity, these games capture many of the crucial problems that arise in the multi-agent setting, such as an exponential number of joint actions or the lack of an explicit coordination mechanism. Our results extend those in Castellini et al. (Proceedings of the 18th International Conference on Autonomous Agents and MultiAgent Systems, AAMAS’19.International Foundation for Autonomous Agents and Multiagent Systems, pp 1862–1864, 2019) and quantify how well various approaches can represent the requisite value functions, and help us identify the reasons that can impede good performance, like sparsity of the values or too tight coordination requirements.