Author:
Haiduk Felix,Zatorre Robert J.,Benjamin Lucas,Morillon Benjamin,Albouy Philippe
Abstract
AbstractSpeech and music are two fundamental modes of human communication. Lateralisation of key processes underlying their perception has been related both to the distinct sensitivity to low-level spectrotemporal acoustic features and to top-down attention. However, the interplay between bottom-up and top-down processes needs to be clarified. In the present study, we investigated the contribution of acoustics and attention to melodies or sentences to lateralisation in fMRI functional network topology. We used sung speech stimuli selectively filtered in temporal or spectral modulation domains with crossed and balanced verbal and melodic content. Perception of speech decreased with degradation of temporal information, whereas perception of melodies decreased with spectral degradation. Applying graph theoretical metrics on fMRI connectivity matrices, we found that local clustering, reflecting functional specialisation, linearly increased when spectral or temporal cues crucial for the task goal were incrementally degraded. These effects occurred in a bilateral fronto-temporo-parietal network for processing temporally degraded sentences and in right auditory regions for processing spectrally degraded melodies. In contrast, global topology remained stable across conditions. These findings suggest that lateralisation for speech and music partially depends on an interplay of acoustic cues and task goals under increased attentional demands.
Funder
Austrian Science Fund
VDS CoBeNe
Università degli Studi di Padova
Canadian Institute for Health Research
Canadian Institute for Advanced Research
ILCB
BLRI
Excellence Initiative of Aix-Marseille University
FRQS
Brain Canada
NSERC
Publisher
Springer Science and Business Media LLC