Sun Kun, Wang Rong
School of Foreign Languages, Tongji University.
Department of Linguistics, University of Tübingen.
Cogn Sci. 2025 Jul;49(7):e70092. doi: 10.1111/cogs.70092.
The majority of research in computational psycholinguistics on sentence processing has focused on word-by-word incremental processing within sentences, rather than holistic sentence-level representations. This study introduces two novel computational approaches for quantifying sentence-level processing: sentence surprisal and sentence relevance. Using multilingual large language models (LLMs), we compute sentence surprisal through three methods, chain rule, next sentence prediction, and negative log-likelihood, and apply a "memory-aware" approach to calculate sentence-level semantic relevance based on convolution operations. The sentence-level metrics developed are tested and compared to validate whether they can predict the reading speed of sentences, and, further, we explore how sentence-level metrics take effects on human processing and comprehending sentences as a whole across languages. The results show that sentence-level metrics are highly capable of predicting sentence reading speed. Our results also indicate that these computational sentence-level metrics are exceptionally effective at predicting and explaining the processing difficulties encountered by readers in processing sentences as a whole across a variety of languages. The proposed sentence-level metrics offer significant interpretability and achieve high accuracy in predicting human sentence reading speed, as they capture unique aspects of comprehension difficulty beyond word-level measures. These metrics serve as valuable computational tools for investigating human sentence processing and advancing our understanding of naturalistic reading. Their strong performance and generalization capabilities highlight their potential to drive progress at the intersection of LLMs and cognitive science.
计算心理语言学中关于句子处理的大多数研究都集中在句子内逐词的增量处理上,而不是整体的句子层面表征。本研究引入了两种用于量化句子层面处理的新颖计算方法:句子意外程度和句子相关性。使用多语言大语言模型(LLMs),我们通过三种方法计算句子意外程度,即链式法则、下一句预测和负对数似然,并应用一种“记忆感知”方法基于卷积运算来计算句子层面的语义相关性。所开发的句子层面指标经过测试和比较,以验证它们是否能够预测句子的阅读速度,此外,我们还探究句子层面指标如何在跨语言的情况下对人类整体处理和理解句子产生影响。结果表明,句子层面指标非常能够预测句子阅读速度。我们的结果还表明,这些计算性句子层面指标在预测和解释读者在跨多种语言整体处理句子时遇到的处理困难方面格外有效。所提出的句子层面指标在预测人类句子阅读速度方面具有显著的可解释性并实现了高精度,因为它们捕捉了超出单词层面度量的理解困难的独特方面。这些指标作为有价值的计算工具,可用于研究人类句子处理并推进我们对自然阅读的理解。它们强大的性能和泛化能力凸显了它们在推动大语言模型与认知科学交叉领域取得进展的潜力。