Searched for: +
(1 - 4 of 4)
document
de Bruin, T.D. (author), Kober, J. (author), Tuyls, Karl (author), Babuska, R. (author)
Deep reinforcement learning makes it possible to train control policies that map high-dimensional observations to actions. These methods typically use gradient-based optimization techniques to enable relatively efficient learning, but are notoriously sensitive to hyperparameter choices and do not have good convergence properties. Gradient...
journal article 2020
document
Alibekov, Eduard (author), Kubalik, Jiri (author), Babuska, R. (author)
This paper addresses the problem of deriving a policy from the value function in the context of critic-only reinforcement learning (RL) in continuous state and action spaces. With continuous-valued states, RL algorithms have to rely on a numerical approximator to represent the value function. Numerical approximation due to its nature virtually...
journal article 2018
document
Verbert, K.A.J. (author), De Schutter, B.H.K. (author), Babuska, R. (author)
Last-minute maintenance planning is often undesirable, as it may cause downtime during operational hours, may require rescheduling of other activities, and does not allow to optimize the management of spare parts, material, and personnel. In spite of the aforementioned drawbacks of last-minute planning, most existing methods plan maintenance...
journal article 2017
document
Najafi, E. (author), Babuska, R. (author), Delgado Lopes, G.A. (author)
Most stabilizing controllers designed for nonlinear systems are valid only within a specific region of the state space, called the domain of attraction (DoA). Computation of the DoA is usually costly and time-consuming. This paper proposes a computationally effective sampling approach to estimate the DoAs of nonlinear systems in real time....
journal article 2016
Searched for: +
(1 - 4 of 4)