Meister Hartmut, Wächtler Moritz, Sandmann Pascale, Lang-Roth Ruth, Abdel-Latif Khaled H A
Department of Otorhinolaryngology, Head and Neck Surgery, University Hospital of Cologne, Kerpenerstr. 62, 50937 Cologne, Germany.
Jean Uhrmacher Institute, University of Cologne, Geibelstraße 29-31, 50931 Cologne, Germany.
Audiol Res. 2025 Jun 24;15(4):77. doi: 10.3390/audiolres15040077.
Sentence stress as part of linguistic prosody plays an important role for verbal communication. It emphasizes particularly important words in a phrase and is reflected by acoustic cues such as the voice fundamental frequency. However, visual cues, especially facial movements, are also important for sentence stress perception. Since cochlear implant (CI) recipients are limited in their use of acoustic prosody cues, the question arises as to what extent they are able to exploit visual features. Virtual characters were used to provide highly realistic but controllable stimuli for investigating sentence stress in groups of experienced CI recipients and typical-hearing (TH) peers. In addition to the proportion of correctly identified stressed words, task load was assessed via reaction times (RTs) and task-evoked pupil dilation (TEPD), and visual attention was estimated via eye tracking. Experiment 1 considered congruent combinations of auditory and visual cues, while Experiment 2 presented incongruent stimuli. In Experiment 1, CI users and TH participants performed similarly in the congruent audiovisual condition, while the former were better at using visual cues. RTs were generally faster in the AV condition, whereas TEPD revealed a more detailed picture, with TH subjects showing greater pupil dilation in the visual condition. The incongruent stimuli in Experiment 2 showed that modality use varied individually among CI recipients, while TH participants relied primarily on auditory cues. Visual cues are generally useful for perceiving sentence stress. As a group, CI users are better at using facial cues than their TH peers. However, CI users show individual differences in the reliability of the various cues.
作为语言韵律一部分的句子重音在言语交流中起着重要作用。它特别强调短语中重要的单词,并通过诸如语音基频等声学线索来体现。然而,视觉线索,尤其是面部动作,对于句子重音感知也很重要。由于人工耳蜗(CI)使用者在利用声学韵律线索方面受到限制,因此就他们能够在多大程度上利用视觉特征这一问题便产生了。虚拟角色被用来提供高度逼真但可控的刺激,以研究有经验的CI使用者组和正常听力(TH)同龄人组中的句子重音。除了正确识别重音单词的比例外,还通过反应时间(RTs)和任务诱发瞳孔扩张(TEPD)来评估任务负荷,并通过眼动追踪来估计视觉注意力。实验1考虑了听觉和视觉线索的一致组合,而实验2呈现了不一致的刺激。在实验1中,CI使用者和TH参与者在一致的视听条件下表现相似,而前者在使用视觉线索方面表现更好。在视听条件下,RTs通常更快,而TEPD揭示了更详细的情况,TH受试者在视觉条件下瞳孔扩张更大。实验2中不一致的刺激表明,CI使用者在模态使用上存在个体差异,而TH参与者主要依赖听觉线索。视觉线索通常有助于感知句子重音。作为一个群体,CI使用者比他们的TH同龄人更善于使用面部线索。然而,CI使用者在各种线索的可靠性方面存在个体差异。