Katz William F, Mehta Sonya
Speech Production Lab, Callier Center for Communication Disorders, School of Behavioral and Brain Sciences, The University of Texas at Dallas Dallas, TX, USA.
Front Hum Neurosci. 2015 Nov 19;9:612. doi: 10.3389/fnhum.2015.00612. eCollection 2015.
Pronunciation training studies have yielded important information concerning the processing of audiovisual (AV) information. Second language (L2) learners show increased reliance on bottom-up, multimodal input for speech perception (compared to monolingual individuals). However, little is known about the role of viewing one's own speech articulation processes during speech training. The current study investigated whether real-time, visual feedback for tongue movement can improve a speaker's learning of non-native speech sounds. An interactive 3D tongue visualization system based on electromagnetic articulography (EMA) was used in a speech training experiment. Native speakers of American English produced a novel speech sound (/ɖ/; a voiced, coronal, palatal stop) before, during, and after trials in which they viewed their own speech movements using the 3D model. Talkers' productions were evaluated using kinematic (tongue-tip spatial positioning) and acoustic (burst spectra) measures. The results indicated a rapid gain in accuracy associated with visual feedback training. The findings are discussed with respect to neural models for multimodal speech processing.
语音训练研究已经得出了有关视听(AV)信息处理的重要信息。第二语言(L2)学习者在语音感知方面表现出对自下而上的多模态输入的依赖增加(与单语个体相比)。然而,对于在语音训练过程中观看自己的语音发音过程所起的作用,人们了解甚少。当前的研究调查了舌头运动的实时视觉反馈是否可以改善说话者对非母语语音的学习。在一项语音训练实验中,使用了基于电磁关节造影(EMA)的交互式3D舌头可视化系统。以美式英语为母语的人在使用3D模型观看自己的语音动作的试验之前、期间和之后发出一种新的语音(/ɖ/;浊音、舌尖龈后腭塞音)。使用运动学(舌尖空间定位)和声学(爆破音频谱)测量方法对说话者的发音进行评估。结果表明,与视觉反馈训练相关的准确性迅速提高。结合多模态语音处理的神经模型对这些发现进行了讨论。