Department of Otorhinolaryngology/Head and Neck Surgery, University Medical Center Groningen, University of Groningen, Groningen, The Netherlands.
CNRS Unité Mixte de Recherche 5292, Lyon Neuroscience Research Center, Auditory Cognition and Psychoacoustics, Institut National de la Santé et de la Recherche Médicale, UMRS 1028, Université Claude Bernard Lyon 1, Université de Lyon, Lyon, France.
J Acoust Soc Am. 2021 Sep;150(3):1620. doi: 10.1121/10.0005938.
Perceptual differences in voice cues, such as fundamental frequency (F0) and vocal tract length (VTL), can facilitate speech understanding in challenging conditions. Yet, we hypothesized that in the presence of spectrotemporal signal degradations, as imposed by cochlear implants (CIs) and vocoders, acoustic cues that overlap for voice perception and phonemic categorization could be mistaken for one another, leading to a strong interaction between linguistic and indexical (talker-specific) content. Fifteen normal-hearing participants performed an odd-one-out adaptive task measuring just-noticeable differences (JNDs) in F0 and VTL. Items used were words (lexical content) or time-reversed words (no lexical content). The use of lexical content was either promoted (by using variable items across comparison intervals) or not (fixed item). Finally, stimuli were presented without or with vocoding. Results showed that JNDs for both F0 and VTL were significantly smaller (better) for non-vocoded compared with vocoded speech and for fixed compared with variable items. Lexical content (forward vs reversed) affected VTL JNDs in the variable item condition, but F0 JNDs only in the non-vocoded, fixed condition. In conclusion, lexical content had a positive top-down effect on VTL perception when acoustic and linguistic variability was present but not on F0 perception. Lexical advantage persisted in the most degraded conditions and vocoding even enhanced the effect of item variability, suggesting that linguistic content could support compensation for poor voice perception in CI users.
感知声音线索(如基频(F0)和声道长度(VTL))的差异可以促进在具有挑战性的条件下的言语理解。然而,我们假设在存在由人工耳蜗(CI)和声码器引起的频谱时间信号退化的情况下,用于语音感知和音位分类的声学线索可能会相互混淆,导致语言和索引(说话人特定)内容之间的强烈相互作用。15 名正常听力参与者执行了一项异常适应任务,测量 F0 和 VTL 的可察觉差异(JND)。使用的项目是单词(词汇内容)或时间反转的单词(无词汇内容)。词汇内容的使用要么是促进的(通过在比较间隔中使用可变项目),要么不是(固定项目)。最后,在没有或有声码的情况下呈现刺激。结果表明,与声码化语音相比,非声码化语音的 F0 和 VTL 的 JND 明显更小(更好),与固定项目相比,可变项目的 JND 更小。词汇内容(正向与反向)在可变项目条件下影响 VTL JND,但仅在非声码化、固定条件下影响 F0 JND。总之,当存在声学和语言可变性时,词汇内容对 VTL 感知有积极的自上而下的影响,但对 F0 感知没有影响。即使在最严重的退化条件和声码化条件下,词汇优势仍然存在,甚至增强了项目可变性的影响,这表明语言内容可以支持补偿 CI 用户较差的语音感知。