Abstract
AbstractUsing biased data to train Artificial Intelligence (AI) algorithms will lead to biased decisions, discriminating against certain groups or individuals. Bias can be explicit (one or several protected features directly influence the decisions) or implicit (one or several protected features indirectly influence the decisions). Unsurprisingly, biased patterns are difficult to detect and mitigate. This paper investigates the extent to which explicit and implicit against one or more protected features in structured classification data sets can be mitigated simultaneously while retaining the data’s discriminatory power. The main contribution of this paper concerns an optimization-based bias mitigation method that reweights the training instances. The algorithm operates with numerical and nominal data and can mitigate implicit and explicit bias against several protected features simultaneously. The trade-off between bias mitigation and accuracy loss can be controlled using parameters in the objective function. The numerical simulations using real-world data sets show a reduction of up to 77% of implicit bias and a complete removal of explicit bias against protected features at no cost of accuracy of a wrapper classifier trained on the data. Overall, the proposed method outperforms the state-of-the-art bias mitigation methods for the selected data sets.
Publisher
Springer Science and Business Media LLC
Reference41 articles.
1. Ahn CW, Ramakrishna R (2003) Elitism-based compact genetic algorithms. IEEE Trans Evol Comput 7(4):367–385. https://doi.org/10.1109/TEVC.2003.814633
2. Balayn A, Lofi C, Houben G-J (2021) Managing bias and unfairness in data for decision support: a survey of machine learning and data engineering approaches to identify and mitigate bias and unfairness within data management and analytics systems. VLDB J 30(5):739–768. https://doi.org/10.1007/s00778-021-00671-8
3. Bellamy RKE, Dey K, Hind M, Hoffman SC, Houde S, Kannan K, Zhang Y (2018) AI Fairness 360: An extensible toolkit for detecting, understanding, and mitigating unwanted algorithmic bias. arXiv:1810.01943
4. Bonchi F, Hajian S, Mishra B, Ramazzotti D (2017) Exposing the probabilistic causal structure of discrimination. Int J Data Sci Anal 3(1):1–21. https://doi.org/10.1007/s41060-016-0040-z
5. Calders T, Kamiran F, Pechenizkiy M (2009) Building classifiers with independency constraints. 2009 IEEE international conference on data mining workshops (pp 13–18)