Author:
Garrido-Muñoz Ismael,Martínez-Santiago Fernando,Montejo-Ráez Arturo
Abstract
AbstractThe study of bias in language models is a growing area of work, however, both research and resources are focused on English. In this paper, we make a first approach focusing on gender bias in some freely available Spanish language models trained using popular deep neural networks, like BERT or RoBERTa. Some of these models are known for achieving state-of-the-art results on downstream tasks. These promising results have promoted such models’ integration in many real-world applications and production environments, which could be detrimental to people affected for those systems. This work proposes an evaluation framework to identify gender bias in masked language models, with explainability in mind to ease the interpretation of the evaluation results. We have evaluated 20 different models for Spanish, including some of the most popular pretrained ones in the research community. Our findings state that varying levels of gender bias are present across these models.This approach compares the adjectives proposed by the model for a set of templates. We classify the given adjectives into understandable categories and compute two new metrics from model predictions, one based on the internal state (probability) and the other one on the external state (rank). Those metrics are used to reveal biased models according to the given categories and quantify the degree of bias of the models under study.
Funder
Plan Nacional I+D+i from the Spanish Government
Andalusian Regional Government
Universidad de Jaén
Publisher
Springer Science and Business Media LLC
Subject
Library and Information Sciences,Linguistics and Language,Education,Language and Linguistics
Reference53 articles.
1. Abdaoui, Amine, & Pradel, Camille. (2020). and Grégoire Sigel. Load What You Need: Smaller Versions of Multilingual BERT. In SustaiNLP / EMNLP.
2. Abid, Abubakar., Farooqi, Maheen., & Zou, James. (2021). Persistent anti-muslim bias in large language models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society, AIES ’21, page 298-306, New York, NY, USA. Association for Computing Machinery. ISBN 9781450384735. https://doi.org/10.1145/3461702.3462624.
3. Al Kuwatly, Hala., Wich, Maximilian., & Groh, Georg. (2020). Identifying and Measuring Annotator Bias Based on Annotators’ Demographic Characteristics. In Proceedings of the Fourth Workshop on Online Abuse and Harms, pages 184–190, Online. Association for Computational Linguistics. https://doi.org/10.18653/v1/2020.alw-1.21.
4. Babaeianjelodar, Marzieh., Lorenz, Stephen., Gordon, Josh., Matthews, Jeanna., & Freitag, Evan. (2020). Quantifying Gender Bias in Different Corpora. In Companion Proceedings of the Web Conference 2020, WWW ’20, page 752-759, New York, NY, USA. Association for Computing Machinery. ISBN 9781450370240. https://doi.org/10.1145/3366424.3383559.
5. Bartl, Marion., Nissim, Malvina., & Gatt, Albert. (2020). Unmasking Contextual Stereotypes: Measuring and Mitigating BERT’s Gender Bias. In Marta R. Costa-jussá, Christian Hardmeier, Kellie Webster, and Will Radford, editors, Proceedings of the Second Workshop on Gender Bias in Natural Language Processing.
Cited by
2 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献