Haro Stephanie, Smalt Christopher J, Ciccarelli Gregory A, Quatieri Thomas F
Human Health and Performance Systems, Massachusetts Institute of Technology Lincoln Laboratory, Lexington, MA, United States.
Speech and Hearing Biosciences and Technology, Harvard Medical School, Boston, MA, United States.
Front Neurosci. 2020 Dec 15;14:588448. doi: 10.3389/fnins.2020.588448. eCollection 2020.
Many individuals struggle to understand speech in listening scenarios that include reverberation and background noise. An individual's ability to understand speech arises from a combination of peripheral auditory function, central auditory function, and general cognitive abilities. The interaction of these factors complicates the prescription of treatment or therapy to improve hearing function. Damage to the auditory periphery can be studied in animals; however, this method alone is not enough to understand the impact of hearing loss on speech perception. Computational auditory models bridge the gap between animal studies and human speech perception. Perturbations to the modeled auditory systems can permit mechanism-based investigations into observed human behavior. In this study, we propose a computational model that accounts for the complex interactions between different hearing damage mechanisms and simulates human speech-in-noise perception. The model performs a digit classification task as a human would, with only acoustic sound pressure as input. Thus, we can use the model's performance as a proxy for human performance. This two-stage model consists of a biophysical cochlear-nerve spike generator followed by a deep neural network (DNN) classifier. We hypothesize that sudden damage to the periphery affects speech perception and that central nervous system adaptation over time may compensate for peripheral hearing damage. Our model achieved human-like performance across signal-to-noise ratios (SNRs) under normal-hearing (NH) cochlear settings, achieving 50% digit recognition accuracy at -20.7 dB SNR. Results were comparable to eight NH participants on the same task who achieved 50% behavioral performance at -22 dB SNR. We also simulated medial olivocochlear reflex (MOCR) and auditory nerve fiber (ANF) loss, which worsened digit-recognition accuracy at lower SNRs compared to higher SNRs. Our simulated performance following ANF loss is consistent with the hypothesis that cochlear synaptopathy impacts communication in background noise more so than in quiet. Following the insult of various cochlear degradations, we implemented extreme and conservative adaptation through the DNN. At the lowest SNRs (<0 dB), both adapted models were unable to fully recover NH performance, even with hundreds of thousands of training samples. This implies a limit on performance recovery following peripheral damage in our human-inspired DNN architecture.
许多人在包含混响和背景噪声的听力场景中难以理解言语。个体理解言语的能力源于外周听觉功能、中枢听觉功能和一般认知能力的综合作用。这些因素的相互作用使得改善听力功能的治疗或疗法的处方变得复杂。听觉外周的损伤可以在动物身上进行研究;然而,仅靠这种方法不足以理解听力损失对言语感知的影响。计算听觉模型弥合了动物研究与人类言语感知之间的差距。对建模的听觉系统的扰动可以允许基于机制的对观察到的人类行为的研究。在本研究中,我们提出了一个计算模型,该模型考虑了不同听力损伤机制之间的复杂相互作用,并模拟了人类在噪声中的言语感知。该模型像人类一样执行数字分类任务,仅将声压作为输入。因此,我们可以将模型的性能用作人类性能的代理。这个两阶段模型由一个生物物理的耳蜗神经尖峰发生器和一个深度神经网络(DNN)分类器组成。我们假设外周的突然损伤会影响言语感知,并且随着时间的推移中枢神经系统的适应可能会补偿外周听力损伤。在正常听力(NH)耳蜗设置下,我们的模型在不同信噪比(SNR)下实现了类似人类的性能,在 -20.7 dB SNR 时实现了 50% 的数字识别准确率。结果与八名在同一任务上的 NH 参与者相当,他们在 -22 dB SNR 时实现了 50% 的行为表现。我们还模拟了内侧橄榄耳蜗反射(MOCR)和听觉神经纤维(ANF)损失,与较高 SNR 相比,这在较低 SNR 时恶化了数字识别准确率。我们模拟的 ANF 损失后的性能与以下假设一致:耳蜗突触病变对背景噪声中的交流影响比对安静环境中的影响更大。在遭受各种耳蜗退化后,我们通过 DNN 实现了极端和保守的适应。在最低 SNR(<0 dB)时,即使有数十万训练样本,两个适应模型都无法完全恢复 NH 性能。这意味着在我们受人类启发的 DNN 架构中外周损伤后性能恢复存在限制。