Influence of Data Processing on the Algorithm Fairness vs. Accuracy Trade-off
Building Pareto Fronts for Equitable Algorithmic Decisions
A.D. Salvi (TU Delft - Electrical Engineering, Mathematics and Computer Science)
S.E. Carter – Mentor (TU Delft - Web Information Systems)
J. Yang – Mentor (TU Delft - Web Information Systems)
Marcus Specht – Graduation committee member (TU Delft - Web Information Systems)
S.N.R. Buijsman – Graduation committee member (TU Delft - Ethics & Philosophy of Technology)
More Info
expand_more
Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.
Abstract
Algorithmic bias due to training from biased data is a widespread issue. Bias mitigation techniques such as fairness-oriented data pre-, in-, and post-processing can help but usually come at the cost of model accuracy. For this contribution, we first conducted a literature review to get a better insight into the potential trade-offs. We followed by implementing a Python program to test how the Disparate Impact Remover (DIR) pre-processing and Reject Option Classification (ROC) post-processing techniques impacted the fairness and accuracy metric values of a Logistic Regression model trained on data from the Adult Income dataset. The implementation also allows for building Pareto fronts that trade off fairness and accuracy metrics of choice, thus offering a blend of perspectives on fairness. Our findings give insight into how combined fairness methods influence the trade-off, but our implementation can be extended to explore such trade-offs using other datasets, models, and fairness methods.