Detecting and Mitigating Algorithmic Bias in Binary Classification using Causal Modeling

Wendy Hui, Wai Kwong Lau

Research output: Chapter in Book/Conference paperConference paperpeer-review

Abstract

This paper proposes the use of causal modeling to detect and mitigate algorithmic bias. We provide a brief description of causal modeling and a general overview of our approach. We then use the Adult dataset, which is available for download from the UC Irvine Machine Learning Repository, to develop (1) a prediction model, which is treated as a black box, and (2) a causal model for bias mitigation. In this paper, we focus on gender bias and the problem of binary classification. We show that gender bias in the prediction model is statistically significant at the 0.05 level. We demonstrate the effectiveness of the causal model in mitigating gender bias by cross-validation. Furthermore, we show that the overall classification accuracy is improved slightly. Our novel approach is intuitive, easy-to-use, and can be implemented using existing statistical software tools such as lavaan in R. Hence, it enhances explainability and promotes trust.

Original languageEnglish
Title of host publicationProceedings - 2024 4th International Conference on Computer Communication and Information Systems, CCCIS 2024
PublisherIEEE, Institute of Electrical and Electronics Engineers
Pages47-51
Number of pages5
ISBN (Print)9798350389548
DOIs
Publication statusPublished - 18 Sept 2024
Event4th International Conference on Computer Communication and Information Systems - Phuket, Thailand
Duration: 27 Feb 202429 Feb 2024

Conference

Conference4th International Conference on Computer Communication and Information Systems
Abbreviated titleCCCIS 2024
Country/TerritoryThailand
CityPhuket
Period27/02/2429/02/24

Fingerprint

Dive into the research topics of 'Detecting and Mitigating Algorithmic Bias in Binary Classification using Causal Modeling'. Together they form a unique fingerprint.

Cite this