Falcon: Fair Active Learning Using Multi-Armed Bandits

Author:

Tae Ki Hyun1,Zhang Hantian2,Park Jaeyoung1,Rong Kexin2,Whang Steven Euijong1

Affiliation:

1. KAIST, Daejeon, Republic of Korea

2. Georgia Institute of Technology, Atlanta, USA

Abstract

Biased data can lead to unfair machine learning models, highlighting the importance of embedding fairness at the beginning of data analysis, particularly during dataset curation and labeling. In response, we propose Falcon, a scalable fair active learning framework. Falcon adopts a data-centric approach that improves machine learning model fairness via strategic sample selection. Given a user-specified group fairness measure, Falcon identifies samples from "target groups" (e.g., (attribute=female, label=positive)) that are the most informative for improving fairness. However, a challenge arises since these target groups are defined using ground truth labels that are not available during sample selection. To handle this, we propose a novel trial-and-error method, where we postpone using a sample if the predicted label is different from the expected one and falls outside the target group. We also observe the trade-off that selecting more informative samples results in higher likelihood of postponing due to undesired label prediction, and the optimal balance varies per dataset. We capture the trade-off between informativeness and postpone rate as policies and propose to automatically select the best policy using adversarial multi-armed bandit methods, given their computational efficiency and theoretical guarantees. Experiments show that Falcon significantly outperforms existing fair active learning approaches in terms of fairness and accuracy and is more efficient. In particular, only Falcon supports a proper trade-off between accuracy and fairness where its maximum fairness score is 1.8--4.5x higher than the second-best results.

Publisher

Association for Computing Machinery (ACM)

Reference69 articles.

1. Naoki Abe and Hiroshi Mamitsuka. 1998. Query Learning Strategies Using Boosting and Bagging. In ICML. San Francisco, CA, USA, 1--9.

2. Jacob Abernethy Pranjal Awasthi Matthäus Kleindessner Jamie Morgenstern Chris Russell and Jie Zhang. 2022. Active Sampling for Min-Max Fairness. In ICML.

3. Jacob D. Abernethy Pranjal Awasthi Matthäus Kleindessner Jamie Morgenstern Chris Russell and Jie Zhang. 2022. Active Sampling for Min-Max Fairness. In ICML. 53--65.

4. Umang Aggarwal Adrian Popescu and Céline Hudelot. 2020. Active Learning for Imbalanced Datasets. In WACV. 1417--1426.

5. Umang Aggarwal Adrian Popescu and Céline Hudelot. 2021. Minority Class Oriented Active Learning for Imbalanced Datasets. In ICPR. 9920--9927.

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3