Lyu Bingjiang, Ge Jianqiao, Niu Zhendong, Tan Li Hai, Gao Jia-Hong
Beijing City Key Laboratory for Medical Physics and Engineering, School of Physics.
Center for MRI Research, Academy for Advanced Interdisciplinary Studies, and.
J Neurosci. 2016 Oct 19;36(42):10813-10822. doi: 10.1523/JNEUROSCI.0583-16.2016.
Spoken language comprehension relies not only on the identification of individual words, but also on the expectations arising from contextual information. A distributed frontotemporal network is known to facilitate the mapping of speech sounds onto their corresponding meanings. However, how prior expectations influence this efficient mapping at the neuroanatomical level, especially in terms of individual words, remains unclear. Using fMRI, we addressed this question in the framework of the dual-stream model by scanning native speakers of Mandarin Chinese, a language highly dependent on context. We found that, within the ventral pathway, the violated expectations elicited stronger activations in the left anterior superior temporal gyrus and the ventral inferior frontal gyrus (IFG) for the phonological-semantic prediction of spoken words. Functional connectivity analysis showed that expectations were mediated by both top-down modulation from the left ventral IFG to the anterior temporal regions and enhanced cross-stream integration through strengthened connections between different subregions of the left IFG. By further investigating the dynamic causality within the dual-stream model, we elucidated how the human brain accomplishes sound-to-meaning mapping for words in a predictive manner.
In daily communication via spoken language, one of the core processes is understanding the words being used. Effortless and efficient information exchange via speech relies not only on the identification of individual spoken words, but also on the contextual information giving rise to expected meanings. Despite the accumulating evidence for the bottom-up perception of auditory input, it is still not fully understood how the top-down modulation is achieved in the extensive frontotemporal cortical network. Here, we provide a comprehensive description of the neural substrates underlying sound-to-meaning mapping and demonstrate how the dual-stream model functions in the modulation of expectations, allowing for a better understanding of how the human brain accomplishes sound-to-meaning mapping in a predictive manner.
口语理解不仅依赖于对单个单词的识别,还依赖于上下文信息产生的预期。已知一个分布式的额颞叶网络有助于将语音映射到其相应的含义。然而,先验预期如何在神经解剖学层面影响这种高效映射,尤其是在单个单词方面,仍不清楚。我们使用功能磁共振成像(fMRI),通过扫描高度依赖上下文的汉语母语者,在双流模型的框架下解决了这个问题。我们发现,在腹侧通路中,对于口语单词的语音-语义预测,违反预期会在左侧颞上回前部和腹侧额下回(IFG)引发更强的激活。功能连接分析表明,预期由从左侧腹侧IFG到颞叶前部区域的自上而下调制以及通过加强左侧IFG不同子区域之间的连接增强跨流整合来介导。通过进一步研究双流模型内的动态因果关系,我们阐明了人类大脑如何以预测方式完成单词的声音到意义的映射。
在通过口语进行的日常交流中,核心过程之一是理解所使用的单词。通过语音进行轻松高效的信息交换不仅依赖于对单个口语单词的识别,还依赖于产生预期含义的上下文信息。尽管有越来越多的证据支持对听觉输入的自下而上感知,但在广泛的额颞叶皮层网络中自上而下调制是如何实现的仍未完全理解。在这里,我们全面描述了声音到意义映射的神经基础,并展示了双流模型在预期调制中的作用,从而更好地理解人类大脑如何以预测方式完成声音到意义的映射。