Abstract
AbstractAs AI technologies are rolled out into healthcare, academia, human resources, law, and a multitude of other domains, they become de-facto arbiters of truth. But truth is highly contested, with many different definitions and approaches. This article discusses the struggle for truth in AI systems and the general responses to date. It then investigates the production of truth in InstructGPT, a large language model, highlighting how data harvesting, model architectures, and social feedback mechanisms weave together disparate understandings of veracity. It conceptualizes this performance as an operationalization of truth, where distinct, often-conflicting claims are smoothly synthesized and confidently presented into truth-statements. We argue that these same logics and inconsistencies play out in Instruct’s successor, ChatGPT, reiterating truth as a non-trivial problem. We suggest that enriching sociality and thickening “reality” are two promising vectors for enhancing the truth-evaluating capacities of future language models. We conclude, however, by stepping back to consider AI truth-telling as a social practice: what kind of “truth” do we as listeners desire?
Funder
The University of Queensland
Publisher
Springer Science and Business Media LLC
Subject
Artificial Intelligence,Human-Computer Interaction,Philosophy
Reference86 articles.
1. Ansari T (2022) “Freaky ChatGPT Fails That Caught Our Eyes!” Analytics India Magazine. https://analyticsindiamag.com/freaky-chatgpt-fails-that-caught-our-eyes/. Accessed 7 Dec 2022
2. Aroyo L, Welty C (2015) Truth is a lie: crowd truth and the seven myths of human annotation. AI Mag 36(1):15–24. https://doi.org/10.1609/aimag.v36i1.2564
3. Askell A, Bai Y, Chen A, Drain D, Ganguli D, Henighan T, Jones A, Joseph N, Mann B, DasSarma N (2021) A general language assistant as a laboratory for alignment. arXiv preprint arXiv:2112.00861
4. Bai Y, Jones A, Ndousse K, Askell A, Chen A, DasSarma N, Drain D et al (2022) Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv:2204.05862
5. Bender EM, Gebru T, McMillan-Major A, Shmitchell S (2021) On the dangers of stochastic parrots: can language models be too big? In: Proceedings of the 2021 ACM conference on fairness, accountability, and transparency, Toronto, Canada. pp 610–23
Cited by
11 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献