Affiliation:
1. Department of Electrical and Computer Engineering, Texas A&M University System, Prairie View A&M University, Prairie View, TX 77446, USA
2. Texas Juvenile Crime Prevention Center, College of Juvenile Justice, Texas A&M University System, Prairie View A&M University, Prairie View, TX 77446, USA
Abstract
Fairness Artificial Intelligence (AI) aims to identify and mitigate bias throughout the AI development process, spanning data collection, modeling, assessment, and deployment—a critical facet of establishing trustworthy AI systems. Tackling data bias through techniques like reweighting samples proves effective for promoting fairness. This paper undertakes a systematic exploration of reweighting samples for conventional Machine-Learning (ML) models, utilizing five models for binary classification on datasets such as Adult Income and COMPAS, incorporating various protected attributes. In particular, AI Fairness 360 (AIF360) from IBM, a versatile open-source library aimed at identifying and mitigating bias in machine-learning models throughout the entire AI application lifecycle, is employed as the foundation for conducting this systematic exploration. The evaluation of prediction outcomes employs five fairness metrics from AIF360, elucidating the nuanced and model-specific efficacy of reweighting samples in fostering fairness within traditional ML frameworks. Experimental results illustrate that reweighting samples effectively reduces bias in traditional ML methods for classification tasks. For instance, after reweighting samples, the balanced accuracy of Decision Tree (DT) improves to 100%, and its bias, as measured by fairness metrics such as Average Odds Difference (AOD), Equal Opportunity Difference (EOD), and Theil Index (TI), is mitigated to 0. However, reweighting samples does not effectively enhance the fairness performance of K Nearest Neighbor (KNN). This sheds light on the intricate dynamics of bias, underscoring the complexity involved in achieving fairness across different models and scenarios.
Reference26 articles.
1. Trustworthy AI: From principles to practices;Li;ACM Comput. Surv.,2023
2. Sharma, S., Zhang, Y., Ríos Aliaga, J.M., Bouneffouf, D., Muthusamy, V., and Varshney, K.R. (2020, January 7–9). Data augmentation for discrimination prevention and bias disambiguation. Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, New York, NY, USA.
3. When algorithmic predictions use human-generated data: A bias-aware classification algorithm for breast cancer diagnosis;Ahsen;Inf. Syst. Res.,2019
4. D-BIAS: A causality-based human-in-the-loop system for tackling algorithmic bias;Ghai;IEEE Trans. Vis. Comput. Graph.,2022
5. Data preprocessing techniques for classification without discrimination;Kamiran;Knowl. Inf. Syst.,2012