Department of Psychological Sciences, University of Connecticut, Storrs, CT, USA.
The Connecticut Institute for the Brain and Cognitive Sciences, Storrs, CT, USA.
Atten Percept Psychophys. 2021 Jul;83(5):2217-2228. doi: 10.3758/s13414-021-02261-w. Epub 2021 Mar 22.
Because different talkers produce their speech sounds differently, listeners benefit from maintaining distinct generative models (sets of beliefs) about the correspondence between acoustic information and phonetic categories for different talkers. A robust literature on phonetic recalibration indicates that when listeners encounter a talker who produces their speech sounds idiosyncratically (e.g., a talker who produces their /s/ sound atypically), they can update their generative model for that talker. Such recalibration has been shown to occur in a relatively talker-specific way. Because listeners in ecological situations often meet several new talkers at once, the present study considered how the process of simultaneously updating two distinct generative models compares to updating one model at a time. Listeners were exposed to two talkers, one who produced /s/ atypically and one who produced /∫/ atypically. Critically, these talkers only produced these sounds in contexts where lexical information disambiguated the phoneme's identity (e.g., epi_ode, flouri_ing). When initial exposure to the two talkers was blocked by voice (Experiment 1), listeners recalibrated to these talkers after relatively little exposure to each talker (32 instances per talker, of which 16 contained ambiguous fricatives). However, when the talkers were intermixed during learning (Experiment 2), listeners required more exposure trials before they were able to adapt to the idiosyncratic productions of these talkers (64 instances per talker, of which 32 contained ambiguous fricatives). Results suggest that there is a perceptual cost to simultaneously updating multiple distinct generative models, potentially because listeners must first select which generative model to update.
由于不同的说话者发出的语音不同,因此听者受益于为不同的说话者保持独特的生成模型(信念集),这种生成模型反映了声学信息与语音类别之间的对应关系。关于语音再校准的大量文献表明,当听者遇到一个以特殊方式发出其语音的说话者(例如,一个以非典型方式发出其/s/音的说话者)时,他们可以更新其针对该说话者的生成模型。这种再校准是以相对特定于说话者的方式发生的。由于在生态情境中,听者通常会同时遇到几个新的说话者,因此本研究考虑了同时更新两个不同的生成模型的过程与一次更新一个模型的过程相比如何。听者接触到两个说话者,一个说话者以非典型方式发出/s/音,另一个说话者以非典型方式发出/∫/音。至关重要的是,这些说话者仅在词汇信息可以消除音位身份歧义的上下文中发出这些音(例如,epi_ode,flouri_ing)。当对两个说话者的初始接触被声音阻断时(实验 1),听者在相对较少地接触每个说话者(每个说话者 32 次,其中 16 次包含模糊的擦音)后就可以对这些说话者进行重新校准。但是,当说话者在学习过程中混合时(实验 2),听者需要更多的接触试验才能适应这些说话者的特殊发音(每个说话者 64 次,其中 32 次包含模糊的擦音)。结果表明,同时更新多个独特的生成模型存在感知成本,这可能是因为听者必须首先选择要更新的生成模型。