Author:
Ashurst Carolyn,Carey Ryan,Chiappa Silvia,Everitt Tom
Abstract
In addition to reproducing discriminatory relationships in the training data, machine learning (ML) systems can also introduce or amplify discriminatory effects. We refer to this as introduced unfairness, and investigate the conditions under which it may arise. To this end, we propose introduced total variation as a measure of introduced unfairness, and establish graphical conditions under which it may be incentivised to occur. These criteria imply that adding the sensitive attribute as a feature removes the incentive for introduced variation under well-behaved loss functions. Additionally, taking a causal perspective, introduced path-specific effects shed light on the issue of when specific paths should be considered fair.
Publisher
Association for the Advancement of Artificial Intelligence (AAAI)
Cited by
4 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. What Fairness Metrics Can Really Tell You: A Case Study in the Educational Domain;Proceedings of the 14th Learning Analytics and Knowledge Conference;2024-03-18
2. Trustworthy Graph Neural Networks: Aspects, Methods, and Trends;Proceedings of the IEEE;2024-02
3. On Imperfect Recall in Multi-Agent Influence Diagrams;Electronic Proceedings in Theoretical Computer Science;2023-07-11
4. Unfair AI: It Isn’t Just Biased Data;2022 IEEE International Conference on Data Mining (ICDM);2022-11