Fair and Private Data Preprocessing through Microaggregation

Author:

González-Zelaya Vladimiro1ORCID,Salas Julián2ORCID,Megías David2ORCID,Missier Paolo3ORCID

Affiliation:

1. Universidad Panamericana, Mexico

2. Universitat Oberta de Catalunya (UOC), Spain

3. Newcastle University, UK

Abstract

Privacy protection for personal data and fairness in automated decisions are fundamental requirements for responsible Machine Learning. Both may be enforced through data preprocessing and share a common target: data should remain useful for a task, while becoming uninformative of the sensitive information. The intrinsic connection between privacy and fairness implies that modifications performed to guarantee one of these goals, may have an effect on the other, e.g., hiding a sensitive attribute from a classification algorithm might prevent a biased decision rule having such attribute as a criterion. This work resides at the intersection of algorithmic fairness and privacy. We show how the two goals are compatible, and may be simultaneously achieved, with a small loss in predictive performance. Our results are competitive with both state-of-the-art fairness correcting algorithms and hybrid privacy-fairness methods. Experiments were performed on three widely used benchmark datasets: Adult Income , COMPAS, and German Credit .

Funder

Ministry of Science and Innovation

CONSENT

SECURING

Publisher

Association for Computing Machinery (ACM)

Subject

General Computer Science

Reference54 articles.

1. Alekh Agarwal Alina Beygelzimer Miroslav Dudík John Langford and Hanna Wallach. 2018. A reductions approach to fair classification. arXiv:1803.02453. Retrieved from https://arxiv.org/abs/1803.02453

2. Arturs Backurs, Piotr Indyk, Krzysztof Onak, Baruch Schieber, Ali Vakilian, and Tal Wagner. 2019. Scalable fair clustering. In Proceedings of the International Conference on Machine Learning. 405–413.

3. Eugene Bagdasaryan, Omid Poursaeed, and Vitaly Shmatikov. 2019. Differential privacy has disparate impact on model accuracy. In Proceedings of the Advances in Neural Information Processing Systems. 15479–15488.

4. Big data’s disparate impact;Barocas Solon;California Law Review,2016

5. Fairness in criminal justice risk assessments: The state-of-the-art;Berk Richard;Sociological Methods and Research,2018

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3