• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

抽样作为句子处理算法理论的合理性。

The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing.

作者信息

Hoover Jacob Louis, Sonderegger Morgan, Piantadosi Steven T, O'Donnell Timothy J

机构信息

McGill University, Montréal, Canada.

Mila Québec AI Institute, Montréal, Canada.

出版信息

Open Mind (Camb). 2023 Jul 21;7:350-391. doi: 10.1162/opmi_a_00086. eCollection 2023.

DOI:10.1162/opmi_a_00086
PMID:37637302
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10449406/
Abstract

Words that are more surprising given context take longer to process. However, no incremental parsing algorithm has been shown to directly predict this phenomenon. In this work, we focus on a class of algorithms whose runtime does naturally scale in surprisal-those that involve repeatedly sampling from the prior. Our first contribution is to show that simple examples of such algorithms predict runtime to increase superlinearly with surprisal, and also predict variance in runtime to increase. These two predictions stand in contrast with literature on surprisal theory (Hale, 2001; Levy, 2008a) which assumes that the expected processing cost increases linearly with surprisal, and makes no prediction about variance. In the second part of this paper, we conduct an empirical study of the relationship between surprisal and reading time, using a collection of modern language models to estimate surprisal. We find that with better language models, reading time increases superlinearly in surprisal, and also that variance increases. These results are consistent with the predictions of sampling-based algorithms.

摘要

在给定语境下更令人惊讶的词汇需要更长时间来处理。然而,尚未有增量解析算法被证明能直接预测这一现象。在这项工作中,我们关注一类算法,其运行时自然地随意外性而扩展,即那些涉及从先验中反复采样的算法。我们的第一个贡献是表明,这类算法的简单示例预测运行时会随意外性超线性增加,并且还预测运行时的方差会增加。这两个预测与意外性理论的文献(黑尔,2001年;利维,2008年a)形成对比,后者假设预期处理成本随意外性线性增加,并且对方差没有预测。在本文的第二部分,我们使用一组现代语言模型来估计意外性,对意外性与阅读时间之间的关系进行了实证研究。我们发现,使用更好的语言模型时,阅读时间随意外性超线性增加,并且方差也增加。这些结果与基于采样的算法的预测一致。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/51673bb71577/opmi-07-350-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/19d6f74ee66f/opmi-07-350-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/cb7996ddd4b6/opmi-07-350-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/d67ce24e6256/opmi-07-350-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/814db4911b67/opmi-07-350-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/f7e8b08c8d9e/opmi-07-350-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/c94993c7727e/opmi-07-350-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/03d74497c5d3/opmi-07-350-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/3532b4a0bd9f/opmi-07-350-i001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/9a70b423d585/opmi-07-350-i002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/59f6590ebc0b/opmi-07-350-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/92735f9844f6/opmi-07-350-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/1e92e35c22b5/opmi-07-350-i003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/f28d3d52589b/opmi-07-350-i004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/8641da2e9c14/opmi-07-350-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/0f079d41c30f/opmi-07-350-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/1e6cf8aab590/opmi-07-350-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/9d4cfcc2a9cd/opmi-07-350-i005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/ae8842d7f810/opmi-07-350-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/51673bb71577/opmi-07-350-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/19d6f74ee66f/opmi-07-350-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/cb7996ddd4b6/opmi-07-350-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/d67ce24e6256/opmi-07-350-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/814db4911b67/opmi-07-350-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/f7e8b08c8d9e/opmi-07-350-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/c94993c7727e/opmi-07-350-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/03d74497c5d3/opmi-07-350-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/3532b4a0bd9f/opmi-07-350-i001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/9a70b423d585/opmi-07-350-i002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/59f6590ebc0b/opmi-07-350-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/92735f9844f6/opmi-07-350-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/1e92e35c22b5/opmi-07-350-i003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/f28d3d52589b/opmi-07-350-i004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/8641da2e9c14/opmi-07-350-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/0f079d41c30f/opmi-07-350-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/1e6cf8aab590/opmi-07-350-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/9d4cfcc2a9cd/opmi-07-350-i005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/ae8842d7f810/opmi-07-350-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4210/10449406/51673bb71577/opmi-07-350-g014.jpg

相似文献

1
The Plausibility of Sampling as an Algorithmic Theory of Sentence Processing.抽样作为句子处理算法理论的合理性。
Open Mind (Camb). 2023 Jul 21;7:350-391. doi: 10.1162/opmi_a_00086. eCollection 2023.
2
Evaluation of an Algorithmic-Level Left-Corner Parsing Account of Surprisal Effects.算法层面左角剖析对惊讶效应的解释评估。
Cogn Sci. 2024 Oct;48(10):e13500. doi: 10.1111/cogs.13500.
3
Comparison of Structural Parsers and Neural Language Models as Surprisal Estimators.作为惊奇度估计器的结构解析器与神经语言模型的比较。
Front Artif Intell. 2022 Mar 3;5:777963. doi: 10.3389/frai.2022.777963. eCollection 2022.
4
Effects of Surprisal and Locality on Danish Sentence Processing: An Eye-Tracking Investigation.意外性和局部性对丹麦语句子处理的影响:一项眼动追踪研究。
J Psycholinguist Res. 2017 Oct;46(5):1119-1136. doi: 10.1007/s10936-017-9482-2.
5
The Influence of Visual Uncertainty on Word Surprisal and Processing Effort.视觉不确定性对单词意外性和加工难度的影响。
Front Psychol. 2018 Dec 14;9:2387. doi: 10.3389/fpsyg.2018.02387. eCollection 2018.
6
Locality and expectation effects in Hindi preverbal constituent ordering.印地语动词前成分语序中的局部性和预期性效应。
Cognition. 2022 Jun;223:104959. doi: 10.1016/j.cognition.2021.104959. Epub 2022 Jan 25.
7
Lossy-Context Surprisal: An Information-Theoretic Model of Memory Effects in Sentence Processing.有损失语境惊讶:句子处理中记忆效应的一种信息论模型。
Cogn Sci. 2020 Mar;44(3):e12814. doi: 10.1111/cogs.12814.
8
Investigating locality effects and surprisal in written English syntactic choice phenomena.探究书面英语句法选择现象中的局部性效应和意外性。
Cognition. 2016 Oct;155:204-232. doi: 10.1016/j.cognition.2016.06.008. Epub 2016 Jul 16.
9
Neurobehavioral Correlates of Surprisal in Language Comprehension: A Neurocomputational Model.语言理解中意外感的神经行为关联:一种神经计算模型。
Front Psychol. 2021 Feb 11;12:615538. doi: 10.3389/fpsyg.2021.615538. eCollection 2021.
10
A framework for modeling the interaction of syntactic processing and eye movement control.句法加工与眼动控制相互作用的建模框架。
Top Cogn Sci. 2013 Jul;5(3):452-74. doi: 10.1111/tops.12026. Epub 2013 May 16.

引用本文的文献

1
Sentence processing by humans and machines: Large language models as a tool to better understand human reading.人类与机器的句子处理:大型语言模型作为更好理解人类阅读的工具
Psychon Bull Rev. 2025 Aug 13. doi: 10.3758/s13423-025-02756-9.
2
EMTeC: A corpus of eye movements on machine-generated texts.EMTeC:机器生成文本上的眼动语料库。
Behav Res Methods. 2025 Jun 3;57(7):189. doi: 10.3758/s13428-025-02677-4.
3
On the Mathematical Relationship Between Contextual Probability and N400 Amplitude.关于情境概率与N400波幅之间的数学关系。

本文引用的文献

1
A Deep Learning Approach to Analyzing Continuous-Time Cognitive Processes.一种用于分析连续时间认知过程的深度学习方法。
Open Mind (Camb). 2024 Mar 13;8:235-264. doi: 10.1162/opmi_a_00126. eCollection 2024.
2
Large-scale evidence for logarithmic effects of word predictability on reading time.大规模证据表明,单词可预测性对阅读时间的影响呈对数关系。
Proc Natl Acad Sci U S A. 2024 Mar 5;121(10):e2307876121. doi: 10.1073/pnas.2307876121. Epub 2024 Feb 29.
3
Comparison of Structural Parsers and Neural Language Models as Surprisal Estimators.
Open Mind (Camb). 2024 Jun 28;8:859-897. doi: 10.1162/opmi_a_00150. eCollection 2024.
4
Word Frequency and Predictability Dissociate in Naturalistic Reading.自然阅读中单词频率与可预测性相互分离。
Open Mind (Camb). 2024 Mar 5;8:177-201. doi: 10.1162/opmi_a_00119. eCollection 2024.
5
Large-scale evidence for logarithmic effects of word predictability on reading time.大规模证据表明,单词可预测性对阅读时间的影响呈对数关系。
Proc Natl Acad Sci U S A. 2024 Mar 5;121(10):e2307876121. doi: 10.1073/pnas.2307876121. Epub 2024 Feb 29.
作为惊奇度估计器的结构解析器与神经语言模型的比较。
Front Artif Intell. 2022 Mar 3;5:777963. doi: 10.3389/frai.2022.777963. eCollection 2022.
4
Language Models Explain Word Reading Times Better Than Empirical Predictability.语言模型比经验可预测性能更好地解释单词阅读时间。
Front Artif Intell. 2022 Feb 2;4:730570. doi: 10.3389/frai.2021.730570. eCollection 2021.
5
The Natural Stories corpus: a reading-time corpus of English texts containing rare syntactic constructions.自然故事语料库:一个包含罕见句法结构的英语文本阅读时间语料库。
Lang Resour Eval. 2021;55(1):63-77. doi: 10.1007/s10579-020-09503-7. Epub 2020 Sep 4.
6
Parsing as a Cue-Based Retrieval Model.句法分析作为一种基于线索的检索模型。
Cogn Sci. 2021 Aug;45(8):e13020. doi: 10.1111/cogs.13020.
7
Continuous-time deconvolutional regression for psycholinguistic modeling.连续时间去卷积回归在心理语言学建模中的应用。
Cognition. 2021 Oct;215:104735. doi: 10.1016/j.cognition.2021.104735. Epub 2021 Jul 21.
8
Single-Stage Prediction Models Do Not Explain the Magnitude of Syntactic Disambiguation Difficulty.单阶段预测模型无法解释句法消歧难度的大小。
Cogn Sci. 2021 Jun;45(6):e12988. doi: 10.1111/cogs.12988.
9
Word predictability effects are linear, not logarithmic: Implications for probabilistic models of sentence comprehension.词汇可预测性效应是线性的,而非对数性的:对句子理解概率模型的启示。
J Mem Lang. 2021 Feb;116. doi: 10.1016/j.jml.2020.104174. Epub 2020 Sep 18.
10
Lossy-Context Surprisal: An Information-Theoretic Model of Memory Effects in Sentence Processing.有损失语境惊讶:句子处理中记忆效应的一种信息论模型。
Cogn Sci. 2020 Mar;44(3):e12814. doi: 10.1111/cogs.12814.