Suppr超能文献

语言模型比经验可预测性能更好地解释单词阅读时间。

Language Models Explain Word Reading Times Better Than Empirical Predictability.

作者信息

Hofmann Markus J, Remus Steffen, Biemann Chris, Radach Ralph, Kuchinke Lars

机构信息

Department of Psychology, University of Wuppertal, Wuppertal, Germany.

Department of Informatics, Universität Hamburg, Hamburg, Germany.

出版信息

Front Artif Intell. 2022 Feb 2;4:730570. doi: 10.3389/frai.2021.730570. eCollection 2021.

Abstract

Though there is a strong consensus that word length and frequency are the most important single-word features determining visual-orthographic access to the mental lexicon, there is less agreement as how to best capture syntactic and semantic factors. The traditional approach in cognitive reading research assumes that word predictability from sentence context is best captured by cloze completion probability (CCP) derived from human performance data. We review recent research suggesting that probabilistic language models provide deeper explanations for syntactic and semantic effects than CCP. Then we compare CCP with three probabilistic language models for predicting word viewing times in an English and a German eye tracking sample: (1) Symbolic n-gram models consolidate syntactic and semantic short-range relations by computing the probability of a word to occur, given two preceding words. (2) Topic models rely on subsymbolic representations to capture long-range semantic similarity by word co-occurrence counts in documents. (3) In recurrent neural networks (RNNs), the subsymbolic units are trained to predict the next word, given all preceding words in the sentences. To examine lexical retrieval, these models were used to predict single fixation durations and gaze durations to capture rapidly successful and standard lexical access, and total viewing time to capture late semantic integration. The linear item-level analyses showed greater correlations of all language models with all eye-movement measures than CCP. Then we examined non-linear relations between the different types of predictability and the reading times using generalized additive models. N-gram and RNN probabilities of the present word more consistently predicted reading performance compared with topic models or CCP. For the effects of last-word probability on current-word viewing times, we obtained the best results with n-gram models. Such count-based models seem to best capture short-range access that is still underway when the eyes move on to the subsequent word. The prediction-trained RNN models, in contrast, better predicted early preprocessing of the next word. In sum, our results demonstrate that the different language models account for differential cognitive processes during reading. We discuss these algorithmically concrete blueprints of lexical consolidation as theoretically deep explanations for human reading.

摘要

尽管人们普遍强烈认为单词长度和频率是决定视觉正字法进入心理词典的最重要的单字特征,但对于如何最好地捕捉句法和语义因素,人们的意见却不太一致。认知阅读研究中的传统方法假定,句子语境中的单词可预测性可以通过从人类表现数据中得出的完形填空完成概率(CCP)来最好地捕捉。我们回顾了最近的研究,这些研究表明概率语言模型比CCP能更深入地解释句法和语义效应。然后我们将CCP与三种概率语言模型进行比较,以预测英语和德语眼动样本中的单词注视时间:(1)符号n元语法模型通过计算给定前两个单词时一个单词出现的概率,巩固句法和语义的短程关系。(2)主题模型依靠亚符号表示,通过文档中的单词共现次数来捕捉长程语义相似性。(3)在循环神经网络(RNN)中,亚符号单元经过训练,根据句子中所有前面的单词来预测下一个单词。为了检验词汇检索,这些模型被用来预测单次注视持续时间和注视持续时间,以捕捉快速成功的和标准的词汇访问,以及总注视时间以捕捉后期语义整合。线性项目级分析表明,所有语言模型与所有眼动指标的相关性都比CCP更高。然后我们使用广义相加模型检验了不同类型的可预测性与阅读时间之间的非线性关系。与主题模型或CCP相比,当前单词的n元语法和RNN概率更一致地预测了阅读表现。对于最后一个单词的概率对当前单词注视时间的影响,我们用n元语法模型得到了最好的结果。这种基于计数的模型似乎能最好地捕捉当眼睛移动到后续单词时仍在进行的短程访问。相比之下,经过预测训练的RNN模型能更好地预测下一个单词的早期预处理。总之,我们的结果表明,不同的语言模型解释了阅读过程中不同的认知过程。我们将这些词汇巩固的算法具体蓝图作为对人类阅读的理论深度解释进行讨论。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e431/8847793/e6cae5821f6c/frai-04-730570-g0001.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验