Searched for: author%3A%22Spaan%2C+M.T.J.%22
(1 - 15 of 15)
document
Spaan, M.T.J. (author), Oliehoek, F.A. (author), Amato, C. (author)
We advance the state of the art in optimal solving of decentralized partially observable Markov decision processes (Dec-POMDPs), which provide a formal model for multiagent planning under uncertainty.
conference paper 2011
document
Spaan, M.T.J. (author), Oliehoek, F.A. (author)
Multiagent Partially Observable Markov Decision Processes (MPOMDPs) provide a powerful framework for optimal decision making under the assumption of instantaneous communication. We focus on a delayed communication setting (MPOMDP-DC), in which broadcasted information is delayed by at most one time step. In this paper, we show that computation of...
conference paper 2012
document
Oliehoek, F.A. (author), Spaan, M.T.J. (author), Amato, C. (author), Whiteson, S. (author)
This article presents the state-of-the-art in optimal solution methods for decentralized partially observable Markov decision processes (Dec-POMDPs), which are general models for collaborative multiagent planning under uncertainty. Building off the generalized multiagent A* (GMAA*) algorithm, which reduces the problem to a tree of one-shot...
journal article 2013
document
Oliehoek, F.A. (author), Whiteson, S. (author), Spaan, M.T.J. (author)
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in...
conference paper 2013
document
Oliehoek, F.A. (author), Whiteson, S. (author), Spaan, M.T.J. (author)
Dec-POMDPs are a powerful framework for planning in multiagent systems, but are provably intractable to solve. This paper proposes a factored forward-sweep policy computation method that tackles the stages of the problem one by one, exploiting weakly coupled structure at each of these stages. An empirical evaluation shows that the loss in...
conference paper 2013
document
Scharpff, J.C.D. (author), Roijers, Diederik M. (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author), de Weerdt, M.M. (author)
In cooperative multi-agent sequential decision making under uncertainty, agents must coordinate to find an optimal joint policy that maximises joint value. Typical algorithms exploit additive structure in the value function, but in the fully-observable multi-agent MDP (MMDP) setting such structure is not present. We propose a new optimal solver...
conference paper 2016
document
Oliehoek, F.A. (author), Spaan, M.T.J. (author), Terwijn, Bas (author), Robbel, Philipp (author), Messias, João V. (author)
This article describes the MultiAgent Decision Process (MADP) toolbox, a software library to support planning and learning for intelligent agents and multiagent systems in uncertain environments. Key features are that it supports partially observable environments and stochastic transition models; has unified support for single- and multiagent...
journal article 2017
document
Satsangi, Yash (author), Whiteson, Shimon (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author)
In active perception tasks, an agent aims to select sensory actions that reduce its uncertainty about one or more hidden variables. For example, a mobile robot takes sensory actions to efficiently navigate in a new environment. While partially observable Markov decision processes (POMDPs) provide a natural model for such problems, reward...
journal article 2018
document
Ponnambalam, C.T. (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author)
The goal in behavior cloning is to extract meaningful information from expertdemonstrations and reproduce the same behavior autonomously. However, theavailable data is unlikely to exhaustively cover the potential problem space. As aresult, the quality of automated decision making is compromised without elegantways to handle the encountering of...
conference paper 2020
document
Ponnambalam, C.T. (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author)
Behavior cloning is a method of automated decision-making that aims to extract meaningful information from expert demonstrations and reproduce the same behavior autonomously. It is unlikely that demonstrations will exhaustively cover the potential problem space, compromising the quality of automation when out-of-distribution states are...
conference paper 2021
document
Smit, Jordi (author), Ponnambalam, C.T. (author), Spaan, M.T.J. (author), Oliehoek, F.A. (author)
Offline reinforcement learning (RL), or learning from a fixed data set, is an attractive alternative to online RL. Offline RL promises to address the cost and safety implications of tak- ing numerous random or bad actions online, a crucial aspect of traditional RL that makes it difficult to apply in real-world problems. However, when RL is na...
conference paper 2021
document
Ponnambalam, C.T. (author), Kamran, Danial (author), Simão, T. D. (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author)
conference paper 2022
document
Suau, M. (author), He, J. (author), Spaan, M.T.J. (author), Oliehoek, F.A. (author)
Learning effective policies for real-world problems is still an open challenge for the field of reinforcement learning (RL). The main limitation being the amount of data needed and the pace at which that data can be obtained. In this paper, we study how to build lightweight simulators of complicated systems that can run sufficiently fast for...
conference paper 2022
document
Suau, M. (author), He, J. (author), Çelikok, Mustafa Mert (author), Spaan, M.T.J. (author), Oliehoek, F.A. (author)
Due to its high sample complexity, simulation is, as of today, critical for the successful application of reinforcement learning. Many real-world problems, however, exhibit overly complex dynamics, which makes their full-scale simulation computationally slow. In this paper, we show how to factorize large networked systems of many agents into...
conference paper 2022
document
Suau, M. (author), He, J. (author), Spaan, M.T.J. (author), Oliehoek, F.A. (author)
Learning effective policies for real-world problems is still an open challenge for the field of reinforcement learning (RL). The main limitation being the amount of data needed and the pace at which that data can be obtained. In this paper, we study how to build lightweight simulators of complicated systems that can run sufficiently fast for...
conference paper 2022
Searched for: author%3A%22Spaan%2C+M.T.J.%22
(1 - 15 of 15)