BACKGROUND
ChatGPT, a new large language model (LLM) developed by OpenAI, has demonstrated impressive performance in several specialized applications. In spite of the rising popularity and performance of artificial intelligence, there are a limited number of studies that evaluate ChatGPT’s capability for multiple choice questions (NCQs) using KIDMAP of Rasch analysis (i.e., an online tool called KIDMAP, which is used to evaluate the performance of ChatGPT in answering MCQs).
OBJECTIVE
Study objectives were to (1) demonstrate the use of online Rasch analysis (namely, RaschOnline) and (2) determine the ChatGPT’s grade compared to a normal sample.
METHODS
ChatGPT capability was evaluated using ten items from Taiwan college entrance examinations for the year 2023. Under a Rasch model, 300 virtual students with normal distributions were simulated and generated to compete with ChatGPR's responses. A total of five visual presentations were created using RaschOnline (e.g., item difficulties, differential item functioning (DIF), item characteristic curve, Wright map, and KIDMAP) to answer the research questions outlined in the objectives.
RESULTS
The results indicated that (1) the difficulty of the ten items monotonously increased from easier to harder (i.e., -2.43, -1.78, -1.48, -0.64, -0.1, 0.33, 0.59, 1.34, 1.7, and 2.47 logits); (2) there was evidence of DIF between gender groups for item 5(p=0.042); (3) item 5 fits the Rasch model rather well (p=0.61); (4) all of the items fit the Rasch model, as indicated by Infit mean square errors below the threshold of 1.5; (5) there was no significant difference in the measures obtained between gender groups(p=0.832); (6) a significant difference was observed among ability grades(P< 0.001); and (7)ChatGPT's capability is graded as A, surpassing grades B to E.
CONCLUSIONS
With RaschOnline, we demonstrate that ChatGPT is capable of scoring a grade A compared to a normal sample, with an excellent ability to answer MCQs of English tests for the year 2023 on Taiwan college entrance examinations.