Loss Bounds for Approximate Influence-Based Abstraction

Conference Paper (2021)
Author(s)

E. Congeduti (TU Delft - Interactive Intelligence)

Alexander Mey (TU Delft - Interactive Intelligence)

Frans Oliehoek (TU Delft - Interactive Intelligence)

Research Group
Interactive Intelligence
Copyright
© 2021 E. Congeduti, A. Mey, F.A. Oliehoek
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 E. Congeduti, A. Mey, F.A. Oliehoek
Research Group
Interactive Intelligence
Pages (from-to)
377-385
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Sequential decision making techniques hold great promise to improve the performance of many real-world systems, but computational complexity hampers their principled application. Influencebased abstraction aims to gain leverage by modeling local subproblems together with the ‘influence’ that the rest of the system exerts on them. While computing exact representations of such influence might be intractable, learning approximate representations offers a promising approach to enable scalable solutions. This paper investigates the performance of such approaches from a theoretical perspective. The primary contribution is the derivation of sufficient conditions on approximate influence representations that can guarantee solutions with small value loss. In particular we show that neural networks trained with cross entropy are well suited to learn approximate influence representations. Moreover, we provide a sample based formulation of the bounds, which reduces the gap to applications. Finally, driven by our theoretical insights, we propose approximation error estimators, which empirically reveal to correlate well with the value loss.

Files

P377.pdf
(pdf | 2.09 Mb)
License info not available