Data-enabled iterative learning control
A zero-sum game design for time-scale-varying tasks
Zhihe Zhuang (Jiangnan University)
Rodrigo A. González (Eindhoven University of Technology)
Hongfeng Tao (Jiangnan University)
Wojciech Paszke (University of Zielona Góra)
Tom Oomen (TU Delft - Team Jan-Willem van Wingerden, Eindhoven University of Technology)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Iterative learning control (ILC) is an intelligent control methodology for tackling iteration-invariant exogenous inputs. It is of great significance to develop its extrapolation for more general repetitive tasks with mutual similarity, e.g., tasks with different time scales. In practice, discrete-time ILC with sampling behavior for time-scale-varying tasks suffers from the failure of perfect corresponding learning and environment-dependent iteration-varying disturbances. This paper develops a novel direct data-based ILC algorithm using off-policy Q-learning for tasks with varying time scales, enabling the robust learning of an optimal ILC policy from experimental input/output (I/O) data. From a two-player zero-sum game perspective, the iteration-varying disturbance generated from the varying time scales of repetitive tasks is tackled quantitatively with a preset disturbance attenuation level. Further, to emphasize the importance of theoretical guarantees of reinforcement learning (RL)-based ILC designs, the data efficiency of the developed algorithm is enhanced based on Willems’ Fundamental Lemma, and a rigorous convergence analysis is given. The simulation model of an F-16 aircraft autopilot is employed to show the effectiveness of the developed approach.
Files
File under embargo until 20-06-2026