• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

扫描模式可预测视觉场景跨模态处理中的句子生成。

Scan patterns predict sentence production in the cross-modal processing of visual scenes.

机构信息

Institute for Language, Cognition and Computation, School of Informatics, University of Edinburgh.

出版信息

Cogn Sci. 2012 Sep-Oct;36(7):1204-23. doi: 10.1111/j.1551-6709.2012.01246.x. Epub 2012 Apr 9.

DOI:10.1111/j.1551-6709.2012.01246.x
PMID:22486717
Abstract

Most everyday tasks involve multiple modalities, which raises the question of how the processing of these modalities is coordinated by the cognitive system. In this paper, we focus on the coordination of visual attention and linguistic processing during speaking. Previous research has shown that objects in a visual scene are fixated before they are mentioned, leading us to hypothesize that the scan pattern of a participant can be used to predict what he or she will say. We test this hypothesis using a data set of cued scene descriptions of photo-realistic scenes. We demonstrate that similar scan patterns are correlated with similar sentences, within and between visual scenes; and that this correlation holds for three phases of the language production process (target identification, sentence planning, and speaking). We also present a simple algorithm that uses scan patterns to accurately predict associated sentences by utilizing similarity-based retrieval.

摘要

大多数日常任务都涉及多种模态,这就提出了一个问题,即认知系统如何协调这些模态的处理。在本文中,我们专注于口语过程中视觉注意和语言处理的协调。先前的研究表明,在视觉场景中的物体被提及之前就已经被注视,这使我们假设参与者的扫描模式可以用来预测他或她会说什么。我们使用一组提示的照片般逼真场景的场景描述数据集来检验这个假设。我们证明,在视觉场景内部和之间,相似的扫描模式与相似的句子相关;并且这种相关性适用于语言产生过程的三个阶段(目标识别、句子规划和说话)。我们还提出了一种简单的算法,该算法通过利用基于相似性的检索,使用扫描模式来准确预测相关句子。

相似文献

1
Scan patterns predict sentence production in the cross-modal processing of visual scenes.扫描模式可预测视觉场景跨模态处理中的句子生成。
Cogn Sci. 2012 Sep-Oct;36(7):1204-23. doi: 10.1111/j.1551-6709.2012.01246.x. Epub 2012 Apr 9.
2
I see what you're saying: the integration of complex speech and scenes during language comprehension.我明白你的意思了:语言理解过程中复杂言语和场景的整合。
Acta Psychol (Amst). 2011 Jun;137(2):208-16. doi: 10.1016/j.actpsy.2011.01.007. Epub 2011 Feb 8.
3
Low-level visual saliency does not predict change detection in natural scenes.低层次视觉显著性无法预测自然场景中的变化检测。
J Vis. 2007 Jul 11;7(10):3.1-10. doi: 10.1167/7.10.3.
4
On the flexibility of grammatical advance planning during sentence production: Effects of cognitive load on multiple lexical access.在句子生成过程中语法提前规划的灵活性:认知负荷对多词汇通达的影响。
J Exp Psychol Learn Mem Cogn. 2010 Mar;36(2):423-40. doi: 10.1037/a0018619.
5
Integrating mechanisms of visual guidance in naturalistic language production.整合自然语言生成中视觉引导的机制。
Cogn Process. 2015 May;16(2):131-50. doi: 10.1007/s10339-014-0642-0. Epub 2014 Nov 23.
6
Fixation and saliency during search of natural scenes: the case of visual agnosia.自然场景搜索过程中的注视与显著性:视觉失认症的案例
Neuropsychologia. 2009 Jul;47(8-9):1994-2003. doi: 10.1016/j.neuropsychologia.2009.03.013. Epub 2009 Mar 18.
7
The role of eye movements in the missing-letter effect revisited with the rapid serial visual presentation procedure.运用快速序列视觉呈现程序对眼动在缺字母效应中的作用进行再探讨。
Can J Exp Psychol. 2010 Mar;64(1):47-52. doi: 10.1037/a0016850.
8
Processing articles and pronouns in agrammatic aphasia: evidence from French.语法缺失性失语症中冠词和代词的处理:来自法语的证据。
Brain Lang. 1994 May;46(4):683-94. doi: 10.1006/brln.1994.1037.
9
Does conspicuity enhance distraction? Saliency and eye landing position when searching for objects.显著性会增加干扰吗?搜索物体时的突出性和眼睛落点位置。
Q J Exp Psychol (Hove). 2009 Jun;62(6):1088-98. doi: 10.1080/17470210802602433. Epub 2009 Jan 10.
10
Salience of the lambs: a test of the saliency map hypothesis with pictures of emotive objects.羔羊的显著性:用情感对象图片对显著性映射假说的一项测试。
J Vis. 2012 Jan 25;12(1):22. doi: 10.1167/12.1.22.

引用本文的文献

1
Individual gaze predicts individual scene descriptions.个体注视可预测个体对场景的描述。
Sci Rep. 2025 Mar 19;15(1):9443. doi: 10.1038/s41598-025-94056-4.
2
Speakers prioritise affordance-based object semantics in scene descriptions.说话者在场景描述中优先考虑基于可供性的物体语义。
Lang Cogn Neurosci. 2023;38(8):1045-1067. doi: 10.1080/23273798.2023.2190136. Epub 2023 Mar 30.
3
Computational framework for fusing eye movements and spoken narratives for image annotation.用于融合眼动与口语叙述以进行图像标注的计算框架。
J Vis. 2020 Jul 1;20(7):13. doi: 10.1167/jov.20.7.13.
4
On Visually-Grounded Reference Production: Testing the Effects of Perceptual Grouping and 2D/3D Presentation Mode.关于视觉基础参照生成:测试知觉分组和二维/三维呈现模式的影响。
Front Psychol. 2019 Oct 1;10:2247. doi: 10.3389/fpsyg.2019.02247. eCollection 2019.
5
Effects of Visual Priming and Event Orientation on Word Order Choice in Russian Sentence Production.视觉启动和事件取向对俄语句子生成中词序选择的影响
Front Psychol. 2019 Aug 20;10:1661. doi: 10.3389/fpsyg.2019.01661. eCollection 2019.
6
Cued by What We See and Hear: Spatial Reference Frame Use in Language.受我们所见所闻的引导:语言中空间参照系的使用
Front Psychol. 2018 Aug 13;9:1287. doi: 10.3389/fpsyg.2018.01287. eCollection 2018.
7
Stored object knowledge and the production of referring expressions: the case of color typicality.存储的物体知识与指代表达的生成:以颜色典型性为例
Front Psychol. 2015 Jul 6;6:935. doi: 10.3389/fpsyg.2015.00935. eCollection 2015.
8
Integrating mechanisms of visual guidance in naturalistic language production.整合自然语言生成中视觉引导的机制。
Cogn Process. 2015 May;16(2):131-50. doi: 10.1007/s10339-014-0642-0. Epub 2014 Nov 23.
9
Sources of variation in developmental language disorders: evidence from eye-tracking studies of sentence production.发展性语言障碍的变异来源:句子产生的眼动研究证据。
Philos Trans R Soc Lond B Biol Sci. 2013 Dec 9;369(1634):20120393. doi: 10.1098/rstb.2012.0393. Print 2014.
10
Where's Wally: the influence of visual salience on referring expression generation.《Where's Wally》:视觉显著性对指称表达生成的影响。
Front Psychol. 2013 Jun 18;4:329. doi: 10.3389/fpsyg.2013.00329. eCollection 2013.