Suppr超能文献

与视觉语音信息对齐。

Alignment to visual speech information.

作者信息

Miller Rachel M, Sanchez Kauyumari, Rosenblum Lawrence D

机构信息

University of California, Riverside, California, USA.

出版信息

Atten Percept Psychophys. 2010 Aug;72(6):1614-25. doi: 10.3758/APP.72.6.1614.

Abstract

Speech alignment is the tendency for interlocutors to unconsciously imitate one another's speaking style. Alignment also occurs when a talker is asked to shadow recorded words (e.g., Shockley, Sabadini, & Fowler, 2004). In two experiments, we examined whether alignment could be induced with visual (lipread) speech and with auditory speech. In Experiment 1, we asked subjects to lipread and shadow out loud a model silently uttering words. The results indicate that shadowed utterances sounded more similar to the model's utterances than did subjects' nonshadowed read utterances. This suggests that speech alignment can be based on visual speech. In Experiment 2, we tested whether raters could perceive alignment across modalities. Raters were asked to judge the relative similarity between a model's visual (silent video) utterance and subjects' audio utterances. The subjects' shadowed utterances were again judged as more similar to the model's than were read utterances, suggesting that raters are sensitive to cross-modal similarity between aligned words.

摘要

言语趋同是指对话者会无意识地相互模仿对方的说话风格。当要求说话者跟读录制的单词时(例如,肖克利、萨巴迪尼和福勒,2004年),也会出现趋同现象。在两项实验中,我们研究了视觉(唇读)言语和听觉言语是否能引发趋同现象。在实验1中,我们要求受试者唇读并大声跟读一个默默说出单词的模型。结果表明,跟读的话语听起来比受试者未跟读的朗读话语更类似于模型的话语。这表明言语趋同可以基于视觉言语。在实验2中,我们测试了评分者是否能察觉到跨模态的趋同现象。要求评分者判断模型的视觉(无声视频)话语与受试者的音频话语之间的相对相似度。受试者的跟读话语再次被判定为比朗读话语更类似于模型的话语,这表明评分者对趋同单词之间的跨模态相似度很敏感。

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验