Abstract
AbstractThe ability to recognize and interpret facial expressions is fundamental to human social cognition, enabling navigation of complex interpersonal interactions and understanding of others’ emotional states. The extent to which neural patterns associated with facial expression processing are shared between observers remains unexplored, and no study has yet examined the neural dynamics specific to different emotional expressions. Additionally, the neural processing dynamics of facial attributes such as sex and identity in relation to facial expressions have not been thoroughly investigated.In this study, we investigated the shared neural dynamics of emotional face processing using an explicit facial emotion recognition task, where participants made two-alternative forced choice (2AFC) decisions on the displayed emotion. Our data-driven approach employed cross-participant multivariate classification and representational dissimilarity analysis on EEG data. The results demonstrate that EEG signals can effectively decode the sex, emotional expression, and identity of face stimuli across different stimuli and participants, indicating shared neural codes for facial expression processing.Multivariate classification analyses revealed that sex is decoded first, followed by identity, and then emotion. Emotional expressions (angry, happy, sad) were decoded earlier when contrasted with neutral expressions. While identity and sex information were modulated by image-level stimulus features, the effects of emotion were independent of visual image properties. Importantly, our findings suggest enhanced processing of face identity and sex for emotional expressions, particularly for angry faces and, to a lesser extent, happy faces.
Publisher
Cold Spring Harbor Laboratory