Searched for: collection%253Air
(1 - 3 of 3)
document
He, K. (author), Shi, S. (author), van den Boom, A.J.J. (author), De Schutter, B.H.K. (author)
Approximate dynamic programming (ADP) faces challenges in dealing with constraints in control problems. Model predictive control (MPC) is, in comparison, well-known for its accommodation of constraints and stability guarantees, although its computation is sometimes prohibitive. This paper introduces an approach combining the two methodologies...
journal article 2024
document
Suau, M. (author), He, J. (author), Congeduti, E. (author), Starre, R.A.N. (author), Czechowski, A.T. (author), Oliehoek, F.A. (author)
Due to its perceptual limitations, an agent may have too little information about the environment to act optimally. In such cases, it is important to keep track of the action-observation history to uncover hidden state information. Recent deep reinforcement learning methods use recurrent neural networks (RNN) to memorize past observations....
journal article 2022
document
Han, Yu (author), Hegyi, A. (author), Zhang, Le (author), He, Zhengbing (author), Chung, Edward (author), Liu, Pan (author)
Conventional reinforcement learning (RL) models of variable speed limit (VSL) control systems (and traffic control systems in general) cannot be trained in real traffic process because new control actions are usually explored randomly, which may result in high costs (delays) due to exploration and learning. For this reason, existing RL-based...
journal article 2022