Affiliation:
1. Department of Psychological and Brain Sciences Villanova University Villanova Pennsylvania USA
2. Psychology Department Gonzaga University Spokane Washington USA
Abstract
AbstractMachine learning techniques have proven to be a useful tool in cognitive neuroscience. However, their implementation in scalp‐recorded electroencephalography (EEG) is relatively limited. To address this, we present three analyses using data from a previous study that examined event‐related potential (ERP) responses to a wide range of naturally‐produced speech sounds. First, we explore which features of the EEG signal best maximize machine learning accuracy for a voicing distinction, using a support vector machine (SVM). We manipulate three dimensions of the EEG signal as input to the SVM: number of trials averaged, number of time points averaged, and polynomial fit. We discuss the trade‐offs in using different feature sets and offer some recommendations for researchers using machine learning. Next, we use SVMs to classify specific pairs of phonemes, finding that we can detect differences in the EEG signal that are not otherwise detectable using conventional ERP analyses. Finally, we characterize the timecourse of phonetic feature decoding across three phonological dimensions (voicing, manner of articulation, and place of articulation), and find that voicing and manner are decodable from neural activity, whereas place of articulation is not. This set of analyses addresses both practical considerations in the application of machine learning to EEG, particularly for speech studies, and also sheds light on current issues regarding the nature of perceptual representations of speech.
Funder
National Science Foundation
Subject
Experimental and Cognitive Psychology,Neuropsychology and Physiological Psychology,Biological Psychiatry,Cognitive Neuroscience,Developmental Neuroscience,Endocrine and Autonomic Systems,Neurology,Experimental and Cognitive Psychology,Neuropsychology and Physiological Psychology,General Neuroscience