Cost Inference for Feedback Dynamic Games from Noisy Partial State Observations and Incomplete Trajectories

Journal Article (2023)
Author(s)

Jingqi Li (University of California)

Chih Yuan Chiu (University of California)

Lasse Peters (TU Delft - Learning & Autonomous Control)

Somayeh Sojoudi (University of California)

Claire J. Tomlin (University of California)

David Fridovich-Keil (The University of Texas at Austin)

Research Group
Learning & Autonomous Control
Copyright
© 2023 Jingqi Li, Chih Yuan Chiu, L. Peters, Somayeh Sojoudi, Claire Tomlin, David Fridovich-Keil
DOI related publication
https://doi.org/10.5555/3545946.3598746
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Jingqi Li, Chih Yuan Chiu, L. Peters, Somayeh Sojoudi, Claire Tomlin, David Fridovich-Keil
Research Group
Learning & Autonomous Control
Pages (from-to)
1062-1070
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

In multi-agent dynamic games, the Nash equilibrium state trajectory of each agent is determined by its cost function and the information pattern of the game. However, the cost and trajectory of each agent may be unavailable to the other agents. Prior work on using partial observations to infer the costs in dynamic games assumes an open-loop information pattern. In this work, we demonstrate that the feedback Nash equilibrium concept is more expressive and encodes more complex behavior. It is desirable to develop specific tools for inferring players' objectives in feedback games. Therefore, we consider the dynamic game cost inference problem under the feedback information pattern, using only partial state observations and incomplete trajectory data. To this end, we first propose an inverse feedback game loss function, whose minimizer yields a feedback Nash equilibrium state trajectory closest to the observation data. We characterize the landscape and differentiability of the loss function. Given the difficulty of obtaining the exact gradient, our main contribution is an efficient gradient approximator, which enables a novel inverse feedback game solver that minimizes the loss using first-order optimization. In thorough empirical evaluations, we demonstrate that our algorithm converges reliably and has better robustness and generalization performance than the open-loop baseline method when the observation data reflects a group of players acting in a feedback Nash game.

Files

3545946.3598746.pdf
(pdf | 1.15 Mb)
- Embargo expired in 30-11-2023
License info not available