McMurray Bob, Farris-Trimble Ashley, Seedorff Michael, Rigler Hannah
1Departments of Psychological and Brain Sciences, Communication Sciences and Disorders, and Linguistics, University of Iowa, Iowa City, Iowa, USA; 2Department of Linguistics, Simon Fraser University, Burnaby, British Columbia, Canada; 3Department of Biostatistics, University of Iowa, Iowa City, Iowa, USA; and 4Department of Psychological and Brain Sciences, University of Iowa, Iowa City, Iowa, USA.
Ear Hear. 2016 Jan-Feb;37(1):e37-51. doi: 10.1097/AUD.0000000000000207.
While outcomes with cochlear implants (CIs) are generally good, performance can be fragile. The authors examined two factors that are crucial for good CI performance. First, while there is a clear benefit for adding residual acoustic hearing to CI stimulation (typically in low frequencies), it is unclear whether this contributes directly to phonetic categorization. Thus, the authors examined perception of voicing (which uses low-frequency acoustic cues) and fricative place of articulation (s/∫, which does not) in CI users with and without residual acoustic hearing. Second, in speech categorization experiments, CI users typically show shallower identification functions. These are typically interpreted as deriving from noisy encoding of the signal. However, psycholinguistic work suggests shallow slopes may also be a useful way to adapt to uncertainty. The authors thus employed an eye-tracking paradigm to examine this in CI users.
Participants were 30 CI users (with a variety of configurations) and 22 age-matched normal hearing (NH) controls. Participants heard tokens from six b/p and six s/∫ continua (eight steps) spanning real words (e.g., beach/peach, sip/ship). Participants selected the picture corresponding to the word they heard from a screen containing four items (a b-, p-, s- and ∫-initial item). Eye movements to each object were monitored as a measure of how strongly they were considering each interpretation in the moments leading up to their final percept.
Mouse-click results (analogous to phoneme identification) for voicing showed a shallower slope for CI users than NH listeners, but no differences between CI users with and without residual acoustic hearing. For fricatives, CI users also showed a shallower slope, but unexpectedly, acoustic + electric listeners showed an even shallower slope. Eye movements showed a gradient response to fine-grained acoustic differences for all listeners. Even considering only trials in which a participant clicked "b" (for example), and accounting for variation in the category boundary, participants made more looks to the competitor ("p") as the voice onset time neared the boundary. CI users showed a similar pattern, but looked to the competitor more than NH listeners, and this was not different at different continuum steps.
Residual acoustic hearing did not improve voicing categorization suggesting it may not help identify these phonetic cues. The fact that acoustic + electric users showed poorer performance on fricatives was unexpected as they usually show a benefit in standardized perception measures, and as sibilants contain little energy in the low-frequency (acoustic) range. The authors hypothesize that these listeners may overweight acoustic input, and have problems when this is not available (in fricatives). Thus, the benefit (or cost) of acoustic hearing for phonetic categorization may be complex. Eye movements suggest that in both CI and NH listeners, phoneme categorization is not a process of mapping continuous cues to discrete categories. Rather listeners preserve gradiency as a way to deal with uncertainty. CI listeners appear to adapt to their implant (in part) by amplifying competitor activation to preserve their flexibility in the face of potential misperceptions.
虽然人工耳蜗(CI)的效果总体良好,但表现可能不稳定。作者研究了对人工耳蜗良好性能至关重要的两个因素。首先,虽然在人工耳蜗刺激中加入残余听觉(通常在低频)有明显益处,但尚不清楚这是否直接有助于语音分类。因此,作者研究了有和没有残余听觉的人工耳蜗使用者对浊音(使用低频声学线索)和擦音发音部位(s/∫,不使用低频声学线索)的感知。其次,在语音分类实验中,人工耳蜗使用者通常表现出较浅的识别函数。这些通常被解释为源于信号的噪声编码。然而,心理语言学研究表明较浅的斜率也可能是适应不确定性的一种有用方式。因此,作者采用眼动追踪范式在人工耳蜗使用者中对此进行研究。
参与者包括30名人工耳蜗使用者(具有多种配置)和22名年龄匹配的正常听力(NH)对照者。参与者听来自六个b/p和六个s/∫连续体(八个步骤)的语音样本,这些语音样本跨越真实单词(例如,beach/peach,sip/ship)。参与者从包含四个项目(一个以b、p、s和∫开头的项目)的屏幕中选择与他们听到的单词相对应的图片。监测对每个物体的眼动,以衡量在形成最终感知之前的瞬间他们对每种解释的考虑程度。
关于浊音的鼠标点击结果(类似于音素识别)显示,人工耳蜗使用者的斜率比正常听力听众浅,但有和没有残余听觉的人工耳蜗使用者之间没有差异。对于擦音,人工耳蜗使用者也表现出较浅的斜率,但出乎意料的是,声电联合使用者表现出更浅的斜率。眼动显示所有听众对细粒度声学差异有梯度反应。例如,即使只考虑参与者点击“b”的试验,并考虑类别边界的变化,随着语音起始时间接近边界,参与者会更多地看向竞争者(“p”)。人工耳蜗使用者表现出类似的模式,但比正常听力听众更多地看向竞争者,并且在不同的连续体步骤中没有差异。
残余听觉并没有改善浊音分类,表明它可能无助于识别这些语音线索。声电联合使用者在擦音上表现较差这一事实出乎意料,因为他们通常在标准化感知测量中表现出优势,而且咝音在低频(声学)范围内能量很少。作者推测这些听众可能过度重视声学输入,并且在没有声学输入时(在擦音中)会出现问题。因此,声学听觉对语音分类的益处(或代价)可能很复杂。眼动表明,在人工耳蜗使用者和正常听力听众中,音素分类都不是将连续线索映射到离散类别的过程。相反,听众保留梯度作为应对不确定性的一种方式。人工耳蜗使用者似乎通过放大竞争者激活来部分适应他们的植入物,以在面对潜在的误感知时保持灵活性。