• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

语言模型在阅读认知模型中优于完形预测能力。

Language models outperform cloze predictability in a cognitive model of reading.

机构信息

Department of Education, Vrije Universiteit Amsterdam, and LEARN! Research Institute, Amsterdam, The Netherlands.

Department of Experimental and Applied Psychology, Vrije Universiteit Amsterdam, Amsterdam, The Netherlands.

出版信息

PLoS Comput Biol. 2024 Sep 25;20(9):e1012117. doi: 10.1371/journal.pcbi.1012117. eCollection 2024 Sep.

DOI:10.1371/journal.pcbi.1012117
PMID:39321153
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11458034/
Abstract

Although word predictability is commonly considered an important factor in reading, sophisticated accounts of predictability in theories of reading are lacking. Computational models of reading traditionally use cloze norming as a proxy of word predictability, but what cloze norms precisely capture remains unclear. This study investigates whether large language models (LLMs) can fill this gap. Contextual predictions are implemented via a novel parallel-graded mechanism, where all predicted words at a given position are pre-activated as a function of contextual certainty, which varies dynamically as text processing unfolds. Through reading simulations with OB1-reader, a cognitive model of word recognition and eye-movement control in reading, we compare the model's fit to eye-movement data when using predictability values derived from a cloze task against those derived from LLMs (GPT-2 and LLaMA). Root Mean Square Error between simulated and human eye movements indicates that LLM predictability provides a better fit than cloze. This is the first study to use LLMs to augment a cognitive model of reading with higher-order language processing while proposing a mechanism on the interplay between word predictability and eye movements.

摘要

虽然词可预测性通常被认为是阅读的一个重要因素,但阅读理论中对可预测性的复杂解释却很缺乏。阅读的计算模型传统上使用 cloze 规范作为词可预测性的代理,但 cloze 规范究竟能捕捉到什么仍然不清楚。本研究探讨了大型语言模型(LLM)是否可以填补这一空白。通过一种新的并行分级机制实现上下文预测,在给定位置的所有预测词都预先激活,作为上下文确定性的函数,上下文确定性随文本处理的展开而动态变化。通过使用 OB1-reader(一种阅读中单词识别和眼动控制的认知模型)进行阅读模拟,我们比较了当使用 cloze 任务得出的可预测性值与从 LLM(GPT-2 和 LLaMA)得出的值来拟合眼动数据时,模型的拟合情况。模拟和人类眼球运动之间的均方根误差表明,LLM 的可预测性提供了比 cloze 更好的拟合。这是第一个使用 LLM 增强阅读认知模型的高阶语言处理的研究,并提出了一个关于词可预测性和眼球运动之间相互作用的机制。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/bbb2b6902252/pcbi.1012117.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/4b798863b66b/pcbi.1012117.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/0a6d15184286/pcbi.1012117.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/3cb1ed84efa7/pcbi.1012117.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/e95330e58938/pcbi.1012117.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/131c49185377/pcbi.1012117.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/bbb2b6902252/pcbi.1012117.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/4b798863b66b/pcbi.1012117.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/0a6d15184286/pcbi.1012117.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/3cb1ed84efa7/pcbi.1012117.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/e95330e58938/pcbi.1012117.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/131c49185377/pcbi.1012117.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f3fb/11458034/bbb2b6902252/pcbi.1012117.g006.jpg

相似文献

1
Language models outperform cloze predictability in a cognitive model of reading.语言模型在阅读认知模型中优于完形预测能力。
PLoS Comput Biol. 2024 Sep 25;20(9):e1012117. doi: 10.1371/journal.pcbi.1012117. eCollection 2024 Sep.
2
Cloze probability, predictability ratings, and computational estimates for 205 English sentences, aligned with existing EEG and reading time data.205 个与现有 EEG 和阅读时间数据对齐的英语句子的 cloze 概率、可预测性评分和计算估计值。
Behav Res Methods. 2024 Aug;56(5):5190-5213. doi: 10.3758/s13428-023-02261-8. Epub 2023 Oct 25.
3
Morphosyntactic but not lexical corpus-based probabilities can substitute for cloze probabilities in reading experiments.基于形态句法而非词汇的语料库概率可以替代 cloze 概率在阅读实验中使用。
PLoS One. 2021 Jan 28;16(1):e0246133. doi: 10.1371/journal.pone.0246133. eCollection 2021.
4
OB1-reader: A model of word recognition and eye movements in text reading.阅读者 OB1:一种文本阅读中单词识别和眼球运动的模型。
Psychol Rev. 2018 Nov;125(6):969-984. doi: 10.1037/rev0000119. Epub 2018 Aug 6.
5
Dissociating word frequency and predictability effects in reading: Evidence from coregistration of eye movements and EEG.阅读中词频与可预测性效应的分离:来自眼动与脑电图同步记录的证据。
J Exp Psychol Learn Mem Cogn. 2015 Nov;41(6):1648-62. doi: 10.1037/xlm0000128. Epub 2015 May 25.
6
Lexical Predictability During Natural Reading: Effects of Surprisal and Entropy Reduction.自然阅读过程中的词汇可预测性:意外性和熵减少的影响。
Cogn Sci. 2018 Jun;42 Suppl 4(Suppl 4):1166-1183. doi: 10.1111/cogs.12597. Epub 2018 Feb 14.
7
Limits on lexical prediction during reading.阅读过程中词汇预测的局限性。
Cogn Psychol. 2016 Aug;88:22-60. doi: 10.1016/j.cogpsych.2016.06.002. Epub 2016 Jul 1.
8
Integrating Large Language Model, EEG, and Eye-Tracking for Word-Level Neural State Classification in Reading Comprehension.将大语言模型、脑电图和眼动追踪相结合,实现阅读理解中单词级神经状态的分类。
IEEE Trans Neural Syst Rehabil Eng. 2024;32:3465-3475. doi: 10.1109/TNSRE.2024.3435460. Epub 2024 Sep 20.
9
Human and computer estimations of Predictability of words in written language.人类和计算机对书面语言中单词可预测性的估计。
Sci Rep. 2020 Mar 10;10(1):4396. doi: 10.1038/s41598-020-61353-z.
10
Language Models Explain Word Reading Times Better Than Empirical Predictability.语言模型比经验可预测性能更好地解释单词阅读时间。
Front Artif Intell. 2022 Feb 2;4:730570. doi: 10.3389/frai.2021.730570. eCollection 2021.

引用本文的文献

1
A systematic evaluation of Dutch large language models' surprisal estimates in sentence, paragraph and book reading.对荷兰大语言模型在句子、段落和书籍阅读中的意外度估计进行的系统评估。
Behav Res Methods. 2025 Aug 18;57(9):266. doi: 10.3758/s13428-025-02774-4.
2
Sentence processing by humans and machines: Large language models as a tool to better understand human reading.人类与机器的句子处理:大型语言模型作为更好理解人类阅读的工具
Psychon Bull Rev. 2025 Aug 13. doi: 10.3758/s13423-025-02756-9.
3
On the spatial limits of parallel word processing in reading.

本文引用的文献

1
Artificial neural networks for model identification and parameter estimation in computational cognitive models.人工神经网络在计算认知模型中的模型识别和参数估计中的应用。
PLoS Comput Biol. 2024 May 15;20(5):e1012119. doi: 10.1371/journal.pcbi.1012119. eCollection 2024 May.
2
Predictive Coding or Just Feature Discovery? An Alternative Account of Why Language Models Fit Brain Data.预测编码还是仅仅是特征发现?关于语言模型为何符合大脑数据的另一种解释。
Neurobiol Lang (Camb). 2024 Apr 1;5(1):64-79. doi: 10.1162/nol_a_00087. eCollection 2024.
3
Large-scale evidence for logarithmic effects of word predictability on reading time.
关于阅读中并行文字处理的空间限制
Atten Percept Psychophys. 2025 Jul;87(5):1704-1716. doi: 10.3758/s13414-025-03101-x. Epub 2025 Jun 5.
大规模证据表明,单词可预测性对阅读时间的影响呈对数关系。
Proc Natl Acad Sci U S A. 2024 Mar 5;121(10):e2307876121. doi: 10.1073/pnas.2307876121. Epub 2024 Feb 29.
4
Prediction during language comprehension: what is next?语言理解中的预测:接下来是什么?
Trends Cogn Sci. 2023 Nov;27(11):1032-1052. doi: 10.1016/j.tics.2023.08.003. Epub 2023 Sep 11.
5
The neuroconnectionist research programme.神经连接主义研究计划。
Nat Rev Neurosci. 2023 Jul;24(7):431-450. doi: 10.1038/s41583-023-00705-w. Epub 2023 May 30.
6
Large Language Models Demonstrate the Potential of Statistical Learning in Language.大型语言模型展示了统计学习在语言中的潜力。
Cogn Sci. 2023 Mar;47(3):e13256. doi: 10.1111/cogs.13256.
7
Prediction as a basis for skilled reading: insights from modern language models.作为熟练阅读基础的预测:来自现代语言模型的见解。
R Soc Open Sci. 2022 Jun 15;9(6):211837. doi: 10.1098/rsos.211837. eCollection 2022 Jun.
8
Shared computational principles for language processing in humans and deep language models.人类和深度语言模型语言处理的共享计算原则。
Nat Neurosci. 2022 Mar;25(3):369-380. doi: 10.1038/s41593-022-01026-4. Epub 2022 Mar 7.
9
Language Models Explain Word Reading Times Better Than Empirical Predictability.语言模型比经验可预测性能更好地解释单词阅读时间。
Front Artif Intell. 2022 Feb 2;4:730570. doi: 10.3389/frai.2021.730570. eCollection 2021.
10
Word predictability effects are linear, not logarithmic: Implications for probabilistic models of sentence comprehension.词汇可预测性效应是线性的,而非对数性的:对句子理解概率模型的启示。
J Mem Lang. 2021 Feb;116. doi: 10.1016/j.jml.2020.104174. Epub 2020 Sep 18.