Emerging trends: When can users trust GPT, and when should they intervene?
-
Published:2024-01-16
Issue:
Volume:
Page:1-11
-
ISSN:1351-3249
-
Container-title:Natural Language Engineering
-
language:en
-
Short-container-title:Nat. Lang. Eng.
Abstract
Abstract
Usage of large language models and chat bots will almost surely continue to grow, since they are so easy to use, and so (incredibly) credible. I would be more comfortable with this reality if we encouraged more evaluations with humans-in-the-loop to come up with a better characterization of when the machine can be trusted and when humans should intervene. This article will describe a homework assignment, where I asked my students to use tools such as chat bots and web search to write a number of essays. Even after considerable discussion in class on hallucinations, many of the essays were full of misinformation that should have been fact-checked. Apparently, it is easier to believe ChatGPT than to be skeptical. Fact-checking and web search are too much trouble.
Publisher
Cambridge University Press (CUP)
Subject
Artificial Intelligence,Linguistics and Language,Language and Linguistics,Software
Reference42 articles.
1. Wang, J. , Hu, X. , Hou, W. , Chen, H. , Zheng, R. , Wang, Y. , Yang, L. , Huang, H. , Ye, W. , Geng, X. , Jiao, B. , Zhang, Y. and Xie, X. (2023). On the robustness of chatgpt: an adversarial and out-of-distribution perspective. ArXiv, abs/2302.12095. 2. Emerging trends: unfair, biased, addictive, dangerous, deadly, and insanely profitable;Church;Natural Language Engineering,2023 3. Morris, J. , Lifland, E. , Yoo, J.Y. , Grigsby, J. , Jin, D. and Qi, Y. (2020). TextAttack: a framework for adversarial attacks, data augmentation, and adversarial training in NLP. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics, pp. 119–126. Online. 4. Chain-of-thought prompting elicits reasoning in large language models;Wei;Advances in Neural Information Processing Systems,2022
|
|