Abstract
AbstractElectrophysiological methods, i.e., M/EEG provide unique views into brain health. Yet, when building predictive models from brain data, it is often unclear how electrophysiology should be combined with other neuroimaging methods. Information can be redundant, useful common representations of multimodal data may not be obvious and multimodal data collection can be medically contraindicated, which reduces applicability. Here, we propose a multimodal model to robustly combine MEG, MRI and fMRI for prediction. We focus on age prediction as surrogate biomarker in 674 subjects from the Cam-CAN. Strikingly, MEG, fMRI and MRI showed additive effects supporting distinct brain-behavior associations. Moreover, the contribution of MEG was best explained by source-topography of power spectra between 8 and 30 Hz. Finally, we demonstrate that the model maintains benefits of stacking when data is missing. The proposed framework hence enables multimodal learning for a wide range of biomarkers from diverse types of brain signals.
Publisher
Cold Spring Harbor Laboratory