Affiliation:
1. Oregon State University, Corvallis, OR, USA
Abstract
The introduction of Large Language Models (LLMs) have captured public imagination and represent a marked improvement in AI in Education (AIED) capabilities. But there is concern that student reliance on automated tools to complete written assignments may lead to a decline in learning. This study investigated whether participant use of LLMs to complete a writing assignment affected retention of learning content. Undergraduate participants ( N = 109) were randomly assigned to complete a writing assignment under one of three conditions: (1), with the assistance of a Retrieval-Augmented Generation (RAG)-based AI psychology tutor; (2) with the assistance of unmodified GPT-4 Turbo; (3) with no AI assistance. After completing the writing task, students completed a posttest quiz to assess their retention of learning material. The control condition had the lowest mean quiz score (M = 9.22, SD = 3.90), followed by the RAG AI tutor condition (M = 10.81, SD = 4.12), and unmodified GPT-4Turbo (M = 11.31, SD = 3.88), with significant differences between the AI tutor condition and the control condition (p = .036); and between the GPT-4 Turbo condition and control (p = .003); but not between the AI tutor and GPT-4 Turbo conditions (p = .283).
Reference20 articles.
1. Baker T., Smith L. (2019). Educ-AI-tion rebooted? Exploring the future of artificial intelligence in schools and colleges. NESTA. https://www.nesta.org.uk/report/education-rebooted/
2. The 2 Sigma Problem: The Search for Methods of Group Instruction as Effective as One-to-One Tutoring
3. Parameter-efficient fine-tuning of large-scale pre-trained language models
4. Dong C. (2023). How to build an AI tutor that can adapt to any course and provide accurate answers using large language model and retrieval-augmented generation. arXiv preprint. https://doi.org/10.48550/arXiv.2311.17696
5. Escape from the Skinner Box: The case for contemporary intelligent learning environments