Author:
Song Jae Yung,Pycha Anne,Culleton Tessa
Abstract
Voice-activated artificially intelligent (voice-AI) assistants, such as Alexa, are remarkably effective at processing spoken commands by native speakers. What happens when the command is produced by an L2 speaker? In the current study, we focused on Korean-speaking L2 learners of English, and we asked (a) whether Alexa could recognize intended productions of two vowel contrasts, /i/ vs. /ɪ/ and /æ/ vs. /ε/, that occur in English but not in Korean, and (b) whether L2 talkers would make clear-speech adjustments when Alexa misrecognized their intended productions. L2 talkers (n = 10) and native English (n = 10) controls asked Alexa to spell out words. Targets were words that formed minimal vowel pairs, e.g., beat-bit, pet-pat. Results showed that Alexa achieved a 55% accuracy rate with L2 productions, compared to 98% for native productions. When Alexa misrecognized an intended production (e.g., spelling P-E-T when the speaker intended pat), L2 talkers adjusted their subsequent production attempts by altering the duration, F1 and F2 of individual vowels (except for /ε/), as well as increasing vowel duration difference between contrasting vowels. These results have implications for theories of speech adaptation, and specifically for our understanding of L2 speech modifications oriented to voice-AI devices.
Subject
Social Sciences (miscellaneous),Communication
Reference54 articles.
1. Perceptual paths to accurate production of L2 vowels: the role of individual differences;Baker;IRAL,2006
2. Child-adult differences in second-language phonological learning: the role of cross-language similarity;Baker;Lang. Speech.,2008
3. Fitting linear mixed-effects models using lme4;Bates;J. Stat. Soft.,2015
4. A direct realist view of cross-language speech perception,;Best,1995
5. BoersmaP.
WeeninkD.
Praat: Doing Phonetics by Computer (Version 6.0.37) [Computer software]2018
Cited by
7 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献