Affiliation:
1. Florida International University, Miami, FL, USA
Abstract
Large Language Models (LLMs) have demonstrated remarkable success across various domains. However, despite their promising performance in numerous real-world applications, most of these algorithms lack fairness considerations. Consequently, they may lead to discriminatory outcomes against certain communities, particularly marginalized populations, prompting extensive study in fair LLMs. On the other hand, fairness in LLMs, in contrast to fairness in traditional machine learning, entails exclusive backgrounds, taxonomies, and fulfillment techniques. To this end, this survey presents a comprehensive overview of recent advances in the existing literature concerning fair LLMs. Specifically, a brief introduction to LLMs is provided, followed by an analysis of factors contributing to bias in LLMs. Additionally, the concept of fairness in LLMs is discussed categorically, summarizing metrics for evaluating bias in LLMs and existing algorithms for promoting fairness. Furthermore, resources for evaluating bias in LLMs, including toolkits and datasets, are summarized. Finally, existing research challenges and open questions are discussed.
Publisher
Association for Computing Machinery (ACM)
Reference175 articles.
1. Persistent Anti-Muslim Bias in Large Language Models
2. Josh Achiam et al. "Gpt-4 technical report". In: arXiv preprint arXiv:2303.08774 (2023).
3. Afra Feyza Aky¨urek et al. "DUnE: Dataset for unified editing". In: arXiv preprint arXiv:2311.16087 (2023).
4. Chantal Amrhein et al. "Exploiting biased models to de-bias text: A gender-fair rewriting model". In: arXiv preprint arXiv:2305.11140 (2023).
5. Haozhe An et al. "Sodapop: open-ended discovery of social biases in social commonsense reasoning models". In: arXiv preprint arXiv:2210.07269 (2022).
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献