Cognitive Network Science Reveals Bias in GPT-3, GPT-3.5 Turbo, and GPT-4 Mirroring Math Anxiety in High-School Students

Author:

Abramski Katherine1,Citraro Salvatore2,Lombardi Luigi3,Rossetti Giulio2ORCID,Stella Massimo3ORCID

Affiliation:

1. Department of Computer Science, University of Pisa, 56127 Pisa, Italy

2. Institute of Information Science and Technologies—National Research Council, 56124 Pisa, Italy

3. Department of Psychology and Cognitive Science, University of Trento, 38122 Trento, Italy

Abstract

Large Language Models (LLMs) are becoming increasingly integrated into our lives. Hence, it is important to understand the biases present in their outputs in order to avoid perpetuating harmful stereotypes, which originate in our own flawed ways of thinking. This challenge requires developing new benchmarks and methods for quantifying affective and semantic bias, keeping in mind that LLMs act as psycho-social mirrors that reflect the views and tendencies that are prevalent in society. One such tendency that has harmful negative effects is the global phenomenon of anxiety toward math and STEM subjects. In this study, we introduce a novel application of network science and cognitive psychology to understand biases towards math and STEM fields in LLMs from ChatGPT, such as GPT-3, GPT-3.5, and GPT-4. Specifically, we use behavioral forma mentis networks (BFMNs) to understand how these LLMs frame math and STEM disciplines in relation to other concepts. We use data obtained by probing the three LLMs in a language generation task that has previously been applied to humans. Our findings indicate that LLMs have negative perceptions of math and STEM fields, associating math with negative concepts in 6 cases out of 10. We observe significant differences across OpenAI’s models: newer versions (i.e., GPT-4) produce 5× semantically richer, more emotionally polarized perceptions with fewer negative associations compared to older versions and N=159 high-school students. These findings suggest that advances in the architecture of LLMs may lead to increasingly less biased models that could even perhaps someday aid in reducing harmful stereotypes in society rather than perpetuating them.

Publisher

MDPI AG

Subject

Artificial Intelligence,Computer Science Applications,Information Systems,Management Information Systems

Reference69 articles.

1. OpenAI (2023). GPT-4 Technical Report. arXiv.

2. Language models are few-shot learners;Brown;Adv. Neural Inf. Process. Syst.,2020

3. Text characterization based on recurrence networks;Silva;Inf. Sci.,2023

4. Using cognitive psychology to understand GPT-3;Binz;Proc. Natl. Acad. Sci. USA,2023

5. Probing the psychology of AI models;Shiffrin;Proc. Natl. Acad. Sci. USA,2023

Cited by 4 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Enhancing Imbalanced Sentiment Analysis: A GPT-3-Based Sentence-by-Sentence Generation Approach;Applied Sciences;2024-01-11

2. A systematic review of ChatGPT use in K‐12 education;European Journal of Education;2023-12-07

3. Using cognitive psychology to understand GPT-like models needs to extend beyond human biases;Proceedings of the National Academy of Sciences;2023-10-16

4. Integrating generative AI in knowledge building;Computers and Education: Artificial Intelligence;2023

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3