Abstract
AbstractIn recent studies, researchers have utilized Large Language Models (LLMs) to investigate semantic representation within the brain. However, in many of these studies, the researchers often examined various semantic information contents separately, such as speech content, objects in scenes, and background stories. To quantitatively evaluate the contribution of various semantic contents in the brain, we recorded brain activity using functional magnetic resonance imaging (fMRI) while participants watched a total of 8.3 hours of videos of dramas or movies. Importantly, we densely annotated these videos at multiple semantic levels related to video contents, which allowed us to extract latent representations of LLMs for a range of semantic contents. We show that LLMs explain human brain activity more accurately than traditional language models, particularly for the high-level background story. Additionally, we show that distinct brain regions correspond to different semantic contents, thereby underscoring the importance of simultaneously modeling various levels of semantic contents. We will make our fMRI dataset publicly available for future research as a biological metric of the alignment between LLMs and humans.
Publisher
Cold Spring Harbor Laboratory
Reference35 articles.
1. A massive 7T fMRI dataset to bridge cognitive neuroscience and artificial intelligence
2. Scaling laws for language encoding models in fmri;arXiv preprint arXiv,2023
3. Discovering event structure in continuous narrative perception and memory;Neuron,2017
4. Caucheteux, C. , Gramfort, A. , and King, J.-R. Disentangling syntax and semantics in the brain with deep net-works. in International conference on machine learning, pp. 1336–1348. PMLR, 2021.
5. Deep language algorithms predict semantic comprehension from brain activity;Scientific reports,2022