Print Email Facebook Twitter Longitudinal Flight Control by Reinforcement Learning Title Longitudinal Flight Control by Reinforcement Learning: Online Adaptive Critic Design Approach to Altitude Control Author Lee, Jun (TU Delft Aerospace Engineering) Contributor van Kampen, Erik-jan (mentor) Degree granting institution Delft University of Technology Programme Aerospace Engineering Date 2019-12-10 Abstract Reinforcement learning is used as a type of adaptive flight control. Adaptive Critic Design (ACD) is a popular approach for online reinforcement learning control due to its explicit generalization of the policy evaluation and the policy improvement elements. A variant of ACD, Incremental Dual Heuristic Programming (IDHP) has previously been developed that allows fully online adaptive control by online identification of system and control matrices. Previous implementation attempts to a high fidelity Cessna Citation model have shown accurate simultaneous altitude and roll angle reference tracking results with outer loop PID and inner loop IDHP rate controllers after an online training phase. This paper presents an implementation attempt to achieve full IDHP altitude control under the influence of measurement noise and atmospheric gusts. Two IDHP controller designs are proposed with and without the cascaded actor structure. Simulation results with measurement noise indicate that the IDHP controller design without the cascaded actor structure can achieve high success ratios. It is demonstrated that IDHP altitude control under measurement noise and atmospheric gusts are achievable under four flight conditions. Subject Reinforcement LearningAdaptive Critic DesignsFlight Control Systems To reference this document use: http://resolver.tudelft.nl/uuid:c1201f27-964c-4257-ad65-89224bef94a1 Part of collection Student theses Document type master thesis Rights © 2019 Jun Lee Files PDF JunHyeonLee_thesis.pdf 13.51 MB Close viewer /islandora/object/uuid:c1201f27-964c-4257-ad65-89224bef94a1/datastream/OBJ/view