On developers’ practices for hazard diagnosis in machine learning systems

Doctoral Thesis (2023)
Author(s)

Agathe Balayn

Contributor(s)

G.J.P.M. Houben – Promotor (TU Delft - Web Information Systems)

A. Bozzon – Promotor (TU Delft - Sustainable Design Engineering, TU Delft - Web Information Systems)

Research Group
Web Information Systems
Copyright
© 2023 A.M.A. Balayn
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 A.M.A. Balayn
Research Group
Web Information Systems
ISBN (electronic)
978-94-6419-926-0
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Machine learning (ML) is an artificial intelligence technology that has a great potential for being adopted in various sectors of activities. Yet, it is now also increasingly recognized as a hazardous technology. Failures in the outputs of an ML system might cause physical or social harms. Besides, the development and deployment of an ML system itself are also argued to be harmful in certain contexts.

Surprisingly, these hazards persist in applications where ML technology has been deployed, despite the increasing amount of research performed by the ML research community. In this thesis, we task ourselves with the challenges of understanding the reasons for the subsistence of hazardous system’s output failures and of hazardous development and deployment processes in practice, and of developing solutions to further diagnose these hazardous failures (especially in the system’s outputs). For that, we investigate further the nature of the potential gap between research and the practices of those developers who build and deploy the systems. To do so, we survey major related ML research directions, surface developers practices and challenges, and search for types of (mis)alignment between theory and practices. There, among others, we find a lack of technical support for ML developers to identify the potential failures of their systems. Hence, we then tackle the development and evaluation of a human-in-the-loop, explainability-based, failure diagnosis method and user-interface for computer vision systems...

Files

License info not available