Easy-read and large language models: on the ethical dimensions of LLM-based text simplification
-
Published:2024-08-04
Issue:3
Volume:26
Page:
-
ISSN:1388-1957
-
Container-title:Ethics and Information Technology
-
language:en
-
Short-container-title:Ethics Inf Technol
Author:
Freyer NilsORCID, Kempt HendrikORCID, Klöser LarsORCID
Abstract
AbstractThe production of easy-read and plain language is a challenging task, requiring well-educated experts to write context-dependent simplifications of texts. Therefore, the domain of easy-read and plain language is currently restricted to the bare minimum of necessary information. Thus, even though there is a tendency to broaden the domain of easy-read and plain language, the inaccessibility of a significant amount of textual information excludes the target audience from partaking or entertainment and restricts their ability to live life autonomously. Large language models can solve a vast variety of natural language tasks, including the simplification of standard language texts to easy-read or plain language. Moreover, with the rise of generative models like GPT, easy-read and plain language may be applicable to all kinds of natural language texts, making formerly inaccessible information accessible to marginalized groups like, a.o., non-native speakers, and people with mental disabilities. In this paper, we argue for the feasibility of text simplification and generation in that context, outline the ethical dimensions, and discuss the implications for researchers in the field of ethics and computer science.
Funder
Bundesministerium für Familie, Senioren, Frauen und Jugend Universitätsklinikum RWTH Aachen
Publisher
Springer Science and Business Media LLC
Reference39 articles.
1. Anderson, E. S. (1999). What is the point of equality? Ethics, 109(2), 287–337. https://doi.org/10.1086/233897 2. Anschütz, M., Oehms, J., Wimmer, T., Jezierski, B., & Groh, G. (2023). Language models for german text simplification: Overcoming parallel data scarcity through style-specific pre-training. https://doi.org/10.48550/ARXIV.2305.12908. 3. Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021, March). On the dangers of stochastic parrots: Can language models be too big?. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610–623). 4. Bhardwaj, R., Majumder, N., & Poria, S. (2021). Investigating gender bias in BERT. Cognitive Computation, 13(4), 1008–1018. https://doi.org/10.1007/s12559-021-09881-2 5. Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., Neelakantan, A., Shyam, P., Sastry, G., Askell, A., Agarwal, S., Herbert-Voss, A., Krueger, G., Henighan, T., Child, R., Ramesh, A., Ziegler, D. M., Jeffrey, Wu., Winter, C., Amodei, D. (2020). Language Models Are Few-Shot Learners.
|
|