Interpretable neural network with limited weights for constructing simple and explainable HI using SHM data

Conference Paper (2022)
Authors

Morteza Moradi (Structural Integrity & Composites)

Panagiotis Komninos (Structural Integrity & Composites)

Benedictus Benedictus (Structural Integrity & Composites)

D. Zarouchas (Structural Integrity & Composites)

Research Group
Structural Integrity & Composites
Copyright
© 2022 M. Moradi, P. Komninos, R. Benedictus, D. Zarouchas
More Info
expand_more
Publication Year
2022
Language
English
Copyright
© 2022 M. Moradi, P. Komninos, R. Benedictus, D. Zarouchas
Research Group
Structural Integrity & Composites
Volume number
14
ISBN (electronic)
9781936263370
DOI:
https://doi.org/10.36001/phmconf.2022.v14i1.3185
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Recently, companies all over the world have been focusing on the improvement of autonomous health management systems in order to enhance performance and reduce downtime costs. To achieve this, the remaining useful life predictions have been given remarkable attention. These predictions depend on the proper designing process and the quality of health indicators (HI) generated from structural health monitoring sensors based on prior established multiple prognostic evaluation criteria. Constructing such HIs from noisy sensory data demands powerful models that enable the automatic selection and fusion of features taken from those relevant measurements. Deep learning models are promising to autonomously extract features in scenarios with a huge volume of data without requiring considerable domain expertise. Nonetheless, the features established by artificial neural networks are complicated to comprehend and cannot be regarded as physical system characteristics. In this regard, the goal of this paper is to extend a new model; an interpretable artificial neural network that enables the automatic selection and fusion of features to construct the most appropriate HIs with remarkably fewer parameters. This model consists of additive and multiplicative layers that provide a feature fusion that better reflects the system’s physical properties. Additionally, the weights are discretized in two ways: a) using a ternary form with values {-1, 0, 1}, and b) relaxing the aforementioned ternary form by rounding the weights at the first decimal point in the range of [-1, 1]. Both discretization techniques have the ability to softly control the number of parameters that should be ignored. This trick guarantees interpretability for the neural network by extracting simple yet powerful equations representing the constructed HIs. Finally, the model’s performance is evaluated and compared with other approaches using a practical case study. The results show that the proposed approach's designed HIs are both interpretable and of high quality according to the criteria of the HI's evaluation.