Aircraft maintenance check scheduling using reinforcement learning

Journal Article (2021)
Author(s)

Pedro Pimenta De Andrade (University of Coimbra, Centre for Informatics and System)

Catarina Ferreira Da Silva (University of Coimbra, Centre for Informatics and System)

Bernardete Ribeiro (University of Coimbra, Centre for Informatics and System)

Bruno Filipe Santos (TU Delft - Air Transport & Operations)

Research Group
Air Transport & Operations
Copyright
© 2021 Pedro Andrade, Catarina Silva, Bernardete Ribeiro, Bruno F. Santos
DOI related publication
https://doi.org/10.3390/aerospace8040113
More Info
expand_more
Publication Year
2021
Language
English
Copyright
© 2021 Pedro Andrade, Catarina Silva, Bernardete Ribeiro, Bruno F. Santos
Research Group
Air Transport & Operations
Issue number
4
Volume number
8
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

This paper presents a Reinforcement Learning (RL) approach to optimize the long-term scheduling of maintenance for an aircraft fleet. The problem considers fleet status, maintenance capacity, and other maintenance constraints to schedule hangar checks for a specified time horizon. The checks are scheduled within an interval, and the goal is to, schedule them as close as possible to their due date. In doing so, the number of checks is reduced, and the fleet availability increases. A Deep Q-learning algorithm is used to optimize the scheduling policy. The model is validated in a real scenario using maintenance data from 45 aircraft. The maintenance plan that is generated with our approach is compared with a previous study, which presented a Dynamic Programming (DP) based approach and airline estimations for the same period. The results show a reduction in the number of checks scheduled, which indicates the potential of RL in solving this problem. The adaptability of RL is also tested by introducing small disturbances in the initial conditions. After training the model with these simulated scenarios, the results show the robustness of the RL approach and its ability to generate efficient maintenance plans in only a few seconds.