Meta AI Research, Paris 75002, France; and Cognitive Neuroimaging Unit NeuroSpin center, 91191, Gif-sur-Yvette, France.
Cognitive Neuroimaging Unit NeuroSpin center, Gif-sur-Yvette, 91191, France.
J Neurosci. 2023 Jul 19;43(29):5350-5364. doi: 10.1523/JNEUROSCI.1163-22.2023. Epub 2023 May 22.
A sentence is more than the sum of its words: its meaning depends on how they combine with one another. The brain mechanisms underlying such semantic composition remain poorly understood. To shed light on the neural vector code underlying semantic composition, we introduce two hypotheses: (1) the intrinsic dimensionality of the space of neural representations should increase as a sentence unfolds, paralleling the growing complexity of its semantic representation; and (2) this progressive integration should be reflected in ramping and sentence-final signals. To test these predictions, we designed a dataset of closely matched normal and jabberwocky sentences (composed of meaningless pseudo words) and displayed them to deep language models and to 11 human participants (5 men and 6 women) monitored with simultaneous MEG and intracranial EEG. In both deep language models and electrophysiological data, we found that representational dimensionality was higher for meaningful sentences than jabberwocky. Furthermore, multivariate decoding of normal versus jabberwocky confirmed three dynamic patterns: (1) a phasic pattern following each word, peaking in temporal and parietal areas; (2) a ramping pattern, characteristic of bilateral inferior and middle frontal gyri; and (3) a sentence-final pattern in left superior frontal gyrus and right orbitofrontal cortex. These results provide a first glimpse into the neural geometry of semantic integration and constrain the search for a neural code of linguistic composition. Starting from general linguistic concepts, we make two sets of predictions in neural signals evoked by reading multiword sentences. First, the intrinsic dimensionality of the representation should grow with additional meaningful words. Second, the neural dynamics should exhibit signatures of encoding, maintaining, and resolving semantic composition. We successfully validated these hypotheses in deep neural language models, artificial neural networks trained on text and performing very well on many natural language processing tasks. Then, using a unique combination of MEG and intracranial electrodes, we recorded high-resolution brain data from human participants while they read a controlled set of sentences. Time-resolved dimensionality analysis showed increasing dimensionality with meaning, and multivariate decoding allowed us to isolate the three dynamical patterns we had hypothesized.
句子的意义不仅仅取决于其组成的单词,还取决于这些单词之间的组合方式。然而,大脑中这种语义组合的机制仍然知之甚少。为了揭示语义组合背后的神经向量代码,我们提出了两个假设:(1)随着句子的展开,神经表示的内在维度应该增加,与语义表示的复杂性增长相平行;(2)这种渐进式整合应该反映在递增和句子结尾的信号中。为了验证这些预测,我们设计了一个由紧密匹配的正常句子和 Jabberwocky 句子(由无意义的伪词组成)组成的数据集,并将其展示给深度语言模型和 11 名同时进行 MEG 和颅内 EEG 监测的人类参与者。在深度语言模型和电生理数据中,我们发现有意义的句子的表示维度高于 Jabberwocky。此外,正常句子与 Jabberwocky 之间的多元解码证实了三种动态模式:(1)每个单词之后的相位模式,在颞叶和顶叶区域达到峰值;(2)一个递增模式,特征在于双侧下额和中额回;(3)在左额上回和右眶额皮质中的句子结尾模式。这些结果提供了语义整合神经几何结构的初步见解,并限制了对语言组合神经代码的搜索。从一般的语言概念出发,我们对阅读多词句子时引发的神经信号做出了两组预测。首先,随着有意义的单词的增加,代表的内在维度应该增加。其次,神经动力学应该表现出编码、保持和解决语义组合的特征。我们在深度神经语言模型中成功验证了这些假设,这些模型是基于文本训练的人工神经网络,在许多自然语言处理任务中表现非常出色。然后,我们使用 MEG 和颅内电极的独特组合,在人类参与者阅读一组受控句子时记录了高分辨率的大脑数据。时间分辨的维度分析显示出随着意义的增加而增加的维度,多元解码使我们能够分离出我们假设的三种动态模式。