Searched for: author%3A%22Whiteson%2C+Shimon%22
(1 - 12 of 12)
document
Li, Guangliang (author), Whiteson, Shimon (author), Dibeklioğlu, Hamdi (author), Hung, H.S. (author)
Interactive reinforcement learning provides a way for agents to learn to solve tasks from evaluative feedback provided by a human user. Previous research showed that humans give copious feedback early in training but very sparsely thereafter. In this paper, we investigate the potential of agent learning from trainers’ facial expressions via...
conference paper 2021
document
Kurin, Vitaly (author), Igl, Maximilian (author), Rocktäschel, Tim (author), Böhmer, J.W. (author), Whiteson, Shimon (author)
Multitask Reinforcement Learning is a promising way to obtain models with better performance, generalisation, data efficiency, and robustness. Most existing work is limited to compatible settings, where the state and action space dimensions are the same across tasks. Graph Neural Networks (GNN) are one way to address incompatible environments,...
conference paper 2021
document
Iqbal, Shariq (author), Witt, Christian A. Schroeder de (author), Peng, Bei (author), Böhmer, J.W. (author), Whiteson, Shimon (author), Sha, Fei (author)
Real world multi-agent tasks often involve varying types and quantities of agents and non-agent entities; however, agents within these tasks rarely need to consider all others at all times in order to act effectively. Factored value function approaches have historically leveraged such independences to improve learning efficiency, but these...
conference paper 2021
document
Gupta, Tarun (author), Mahajan, Anuj (author), Peng, Bei (author), Böhmer, J.W. (author), Whiteson, Shimon (author)
VDN and QMIX are two popular value-based algorithms for cooperative MARL that learn a centralized action value function as a monotonic mixing of per-agent utilities. While this enables easy decentralization of the learned policy, the restricted joint action value function can prevent them from solving tasks that require significant coordination...
conference paper 2021
document
Igl, Maximilian (author), Farquhar, Gregory (author), Luketina, Jelena (author), Böhmer, J.W. (author), Whiteson, Shimon (author)
Non-stationarity can arise in Reinforcement Learning (RL) even in stationary environments. For example, most RL algorithms collect new data throughout training, using a non-stationary behaviour policy. Due to the transience of this non-stationarity, it is often not explicitly addressed in deep RL and a single neural network is continually...
conference paper 2021
document
Castellini, Jacopo (author), Oliehoek, F.A. (author), Savani, Rahul (author), Whiteson, Shimon (author)
Recent years have seen the application of deep reinforcement learning techniques to cooperative multi-agent systems, with great empirical success. However, given the lack of theoretical insight, it remains unclear what the employed neural networks are learning, or how we should enhance their learning power to address the problems on which...
journal article 2021
document
Satsangi, Yash (author), Lim, Sungsu (author), Whiteson, Shimon (author), Oliehoek, F.A. (author), White, Martha (author)
Information gathering in a partially observable environment can be formulated as a reinforcement learning (RL), problem where the reward depends on the agent's uncertainty. For example, the reward can be the negative entropy of the agent's belief over an unknown (or hidden) variable. Typically, the rewards of an RL agent are defined as a...
conference paper 2020
document
Igl, Maximilian (author), Gambardella, Andrew (author), He, J. (author), Nardelli, Nantas (author), Siddharth, N (author), Böhmer, J.W. (author), Whiteson, Shimon (author)
We present Multitask Soft Option Learning (MSOL), a hierarchical multitask framework based on Planning as Inference. MSOL extends the concept of options, using separate variational posteriors for each task, regularized by a shared prior. This “soft” version of options avoids several instabilities during training in a multitask setting, and...
conference paper 2020
document
Li, Guangliang (author), Dibeklioğlu, Hamdi (author), Whiteson, Shimon (author), Hung, H.S. (author)
Interactive reinforcement learning provides a way for agents to learn to solve tasks from evaluative feedback provided by a human user. Previous research showed that humans give copious feedback early in training but very sparsely thereafter. In this article, we investigate the potential of agent learning from trainers’ facial expressions via...
journal article 2020
document
Castellini, Jacopo (author), Oliehoek, F.A. (author), Savani, Rahul (author), Whiteson, Shimon (author)
Recent years have seen the application of deep reinforcement learning techniques to cooperative multi-agent systems, with great empirical success. In this work, we empirically investigate the representational power of various network architectures on a series of one-shot games. Despite their simplicity, these games capture many of the crucial...
conference paper 2019
document
Satsangi, Yash (author), Whiteson, Shimon (author), Oliehoek, F.A. (author), Spaan, M.T.J. (author)
In active perception tasks, an agent aims to select sensory actions that reduce its uncertainty about one or more hidden variables. For example, a mobile robot takes sensory actions to efficiently navigate in a new environment. While partially observable Markov decision processes (POMDPs) provide a natural model for such problems, reward...
journal article 2018
document
Li, Guangliang (author), Whiteson, Shimon (author), Bradley Knox, W (author), Hung, H.S. (author)
Learning from rewards generated by a human trainer observing an agent in action has been proven to be a powerful method for teaching autonomous agents to perform challenging tasks, especially for those non-technical users. Since the efficacy of this approach depends critically on the reward the trainer provides, we consider how the...
journal article 2018
Searched for: author%3A%22Whiteson%2C+Shimon%22
(1 - 12 of 12)