Abrams Ellie Bean, Marantz Alec, Krementsov Isaac, Gwilliams Laura
Department of Psychology, New York University, New York, New York 10003
Center for Language, Music, and Emotion (CLaME), New York University, New York, New York 10003.
J Neurosci. 2025 Mar 19;45(12):e1111242025. doi: 10.1523/JNEUROSCI.1111-24.2025.
The ability to perceive pitch allows human listeners to experience music, recognize the identity and emotion conveyed by conversational partners, and make sense of their auditory environment. A pitch percept is formed by weighting different acoustic cues (e.g., signal fundamental frequency and interharmonic spacing) and contextual cues (expectation). How and when such cues are neurally encoded and integrated remains debated. In this study, 28 participants (16 female) listened to tone sequences with different acoustic cues (pure tones, complex missing fundamental tones, and tones with an ambiguous mixture), placed in predictable and less predictable sequences, while magnetoencephalography was recorded. Decoding analyses revealed that pitch was encoded in neural responses to all three tone types in the low-to-mid auditory cortex and sensorimotor cortex bilaterally, with right-hemisphere dominance. The pattern of activity generalized across cue types, offset in time: pitch was neurally encoded earlier for harmonic tones (∼85 ms) than pure tones (∼95 ms). For ambiguous tones, pitch emerged significantly earlier in predictable contexts than in unpredictable. The results suggest that a unified neural representation of pitch emerges by integrating independent pitch cues and that context alters the dynamics of pitch generation when acoustic cues are ambiguous.
感知音高的能力使人类听众能够体验音乐、识别对话伙伴传达的身份和情感,并理解他们的听觉环境。音高感知是通过权衡不同的声学线索(例如信号基频和谐波间间距)以及情境线索(期望)而形成的。这些线索如何以及何时在神经上进行编码和整合仍存在争议。在这项研究中,28名参与者(16名女性)聆听了具有不同声学线索的音调序列(纯音、复杂的缺失基音以及具有模糊混合的音调),这些序列被放置在可预测和较不可预测的序列中,同时记录了脑磁图。解码分析表明,音高在双侧低至中听觉皮层和感觉运动皮层中对所有三种音调类型的神经反应中进行编码,以右侧半球占主导。活动模式在不同线索类型中具有普遍性,但在时间上有所偏移:对于谐波音调,音高在神经上的编码时间(约85毫秒)早于纯音(约95毫秒)。对于模糊音调,在可预测情境中,音高出现的时间明显早于不可预测情境。结果表明,通过整合独立的音高线索,形成了统一的音高神经表征,并且当声学线索模糊时,情境会改变音高产生的动态过程。