Suppr超能文献

维度和渐变:大脑和深度语言模型动态中句子整合的特征。

Dimensionality and Ramping: Signatures of Sentence Integration in the Dynamics of Brains and Deep Language Models.

机构信息

Meta AI Research, Paris 75002, France; and Cognitive Neuroimaging Unit NeuroSpin center, 91191, Gif-sur-Yvette, France.

Cognitive Neuroimaging Unit NeuroSpin center, Gif-sur-Yvette, 91191, France.

出版信息

J Neurosci. 2023 Jul 19;43(29):5350-5364. doi: 10.1523/JNEUROSCI.1163-22.2023. Epub 2023 May 22.

Abstract

A sentence is more than the sum of its words: its meaning depends on how they combine with one another. The brain mechanisms underlying such semantic composition remain poorly understood. To shed light on the neural vector code underlying semantic composition, we introduce two hypotheses: (1) the intrinsic dimensionality of the space of neural representations should increase as a sentence unfolds, paralleling the growing complexity of its semantic representation; and (2) this progressive integration should be reflected in ramping and sentence-final signals. To test these predictions, we designed a dataset of closely matched normal and jabberwocky sentences (composed of meaningless pseudo words) and displayed them to deep language models and to 11 human participants (5 men and 6 women) monitored with simultaneous MEG and intracranial EEG. In both deep language models and electrophysiological data, we found that representational dimensionality was higher for meaningful sentences than jabberwocky. Furthermore, multivariate decoding of normal versus jabberwocky confirmed three dynamic patterns: (1) a phasic pattern following each word, peaking in temporal and parietal areas; (2) a ramping pattern, characteristic of bilateral inferior and middle frontal gyri; and (3) a sentence-final pattern in left superior frontal gyrus and right orbitofrontal cortex. These results provide a first glimpse into the neural geometry of semantic integration and constrain the search for a neural code of linguistic composition. Starting from general linguistic concepts, we make two sets of predictions in neural signals evoked by reading multiword sentences. First, the intrinsic dimensionality of the representation should grow with additional meaningful words. Second, the neural dynamics should exhibit signatures of encoding, maintaining, and resolving semantic composition. We successfully validated these hypotheses in deep neural language models, artificial neural networks trained on text and performing very well on many natural language processing tasks. Then, using a unique combination of MEG and intracranial electrodes, we recorded high-resolution brain data from human participants while they read a controlled set of sentences. Time-resolved dimensionality analysis showed increasing dimensionality with meaning, and multivariate decoding allowed us to isolate the three dynamical patterns we had hypothesized.

摘要

句子的意义不仅仅取决于其组成的单词,还取决于这些单词之间的组合方式。然而,大脑中这种语义组合的机制仍然知之甚少。为了揭示语义组合背后的神经向量代码,我们提出了两个假设:(1)随着句子的展开,神经表示的内在维度应该增加,与语义表示的复杂性增长相平行;(2)这种渐进式整合应该反映在递增和句子结尾的信号中。为了验证这些预测,我们设计了一个由紧密匹配的正常句子和 Jabberwocky 句子(由无意义的伪词组成)组成的数据集,并将其展示给深度语言模型和 11 名同时进行 MEG 和颅内 EEG 监测的人类参与者。在深度语言模型和电生理数据中,我们发现有意义的句子的表示维度高于 Jabberwocky。此外,正常句子与 Jabberwocky 之间的多元解码证实了三种动态模式:(1)每个单词之后的相位模式,在颞叶和顶叶区域达到峰值;(2)一个递增模式,特征在于双侧下额和中额回;(3)在左额上回和右眶额皮质中的句子结尾模式。这些结果提供了语义整合神经几何结构的初步见解,并限制了对语言组合神经代码的搜索。从一般的语言概念出发,我们对阅读多词句子时引发的神经信号做出了两组预测。首先,随着有意义的单词的增加,代表的内在维度应该增加。其次,神经动力学应该表现出编码、保持和解决语义组合的特征。我们在深度神经语言模型中成功验证了这些假设,这些模型是基于文本训练的人工神经网络,在许多自然语言处理任务中表现非常出色。然后,我们使用 MEG 和颅内电极的独特组合,在人类参与者阅读一组受控句子时记录了高分辨率的大脑数据。时间分辨的维度分析显示出随着意义的增加而增加的维度,多元解码使我们能够分离出我们假设的三种动态模式。

相似文献

6
Neural Encoding and Decoding With Distributed Sentence Representations.分布式句子表示的神经编码和解码。
IEEE Trans Neural Netw Learn Syst. 2021 Feb;32(2):589-603. doi: 10.1109/TNNLS.2020.3027595. Epub 2021 Feb 4.
8
An Integrated Neural Decoder of Linguistic and Experiential Meaning.语言和体验意义的综合神经解码器。
J Neurosci. 2019 Nov 6;39(45):8969-8987. doi: 10.1523/JNEUROSCI.2575-18.2019. Epub 2019 Sep 30.

引用本文的文献

6
Language is primarily a tool for communication rather than thought.语言主要是一种交流工具,而不是思维工具。
Nature. 2024 Jun;630(8017):575-586. doi: 10.1038/s41586-024-07522-w. Epub 2024 Jun 19.

本文引用的文献

3
Biwhitening Reveals the Rank of a Count Matrix.双白化揭示计数矩阵的秩。
SIAM J Math Data Sci. 2022;4(4):1420-1446. doi: 10.1137/21m1456807.
4
Neural representational geometry underlies few-shot concept learning.神经表象几何是少样本概念学习的基础。
Proc Natl Acad Sci U S A. 2022 Oct 25;119(43):e2200800119. doi: 10.1073/pnas.2200800119. Epub 2022 Oct 17.
5
A hierarchy of linguistic predictions during natural language comprehension.自然语言理解过程中的语言预测层次。
Proc Natl Acad Sci U S A. 2022 Aug 9;119(32):e2201968119. doi: 10.1073/pnas.2201968119. Epub 2022 Aug 3.
10
The population doctrine in cognitive neuroscience.认知神经科学中的群体学说。
Neuron. 2021 Oct 6;109(19):3055-3068. doi: 10.1016/j.neuron.2021.07.011. Epub 2021 Aug 19.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验