“☑ Fairness Toolkits, A Checkbox Culture?” On the Factors that Fragment Developer Practices in Handling Algorithmic Harms

Conference Paper (2023)
Author(s)

A.M.A. Balayn (TU Delft - Web Information Systems)

M. Yurrita Semperena (TU Delft - Human Technology Relations)

Jie Yang (TU Delft - Web Information Systems)

Ujwal Gadiraju (TU Delft - Web Information Systems)

Research Group
Web Information Systems
Copyright
© 2023 A.M.A. Balayn, M. Yurrita Semperena, J. Yang, Ujwal Gadiraju
DOI related publication
https://doi.org/10.1145/3600211.3604674
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 A.M.A. Balayn, M. Yurrita Semperena, J. Yang, Ujwal Gadiraju
Research Group
Web Information Systems
Pages (from-to)
482–495
ISBN (print)
979-8-4007-0231-0
ISBN (electronic)
9798400702310
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Fairness toolkits are developed to support machine learning (ML) practitioners in using algorithmic fairness metrics and mitigation methods. Past studies have investigated practical challenges for toolkit usage, which are crucial to understanding how to support practitioners. However, the extent to which fairness toolkits impact practitioners’ practices and enable reflexivity around algorithmic harms remains unclear (i.e., distributive unfairness beyond algorithmic fairness, and harms that are not related to the outputs of ML systems). Little is currently understood about the root factors that fragment practices when using fairness toolkits and how practitioners reflect on algorithmic harms. Yet, a deeper understanding of these facets is essential to enable the design of support tools for practitioners. To investigate the impact of toolkits on practices and identify factors that shape these practices, we carried out a qualitative study with 30 ML practitioners with varying backgrounds. Through a mixed within and between-subjects design, we tasked the practitioners with developing an ML model, and analyzed their reported practices to surface potential factors that lead to differences in practices. Interestingly, we found that fairness toolkits act as double-edge swords — with potentially positive and negative impacts on practices. Our findings showcase a plethora of human and organizational factors that play a key role in the way toolkits are envisioned and employed. These results bear implications for the design of future toolkits and educational training for practitioners and call for the creation of new policies to handle the organizational constraints faced by practitioners.