• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过主动高效编码实现主动双目视觉和运动视觉的自主发展。

Autonomous Development of Active Binocular and Motion Vision Through Active Efficient Coding.

作者信息

Lelais Alexander, Mahn Jonas, Narayan Vikram, Zhang Chong, Shi Bertram E, Triesch Jochen

机构信息

Frankfurt Institute for Advanced Studies, Frankfurt, Germany.

Department of Electronic and Computer Engineering, Hong Kong University of Science and Technology, Kowloon, Hong Kong.

出版信息

Front Neurorobot. 2019 Jul 16;13:49. doi: 10.3389/fnbot.2019.00049. eCollection 2019.

DOI:10.3389/fnbot.2019.00049
PMID:31379548
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC6646586/
Abstract

We present a model for the autonomous and simultaneous learning of active binocular and motion vision. The model is based on the Active Efficient Coding (AEC) framework, a recent generalization of classic efficient coding theories to active perception. The model learns how to efficiently encode the incoming visual signals generated by an object moving in 3-D through sparse coding. Simultaneously, it learns how to produce eye movements that further improve the efficiency of the sensory coding. This learning is driven by an intrinsic motivation to maximize the system's coding efficiency. We test our approach on the humanoid robot iCub using simulations. The model demonstrates self-calibration of accurate object fixation and tracking of moving objects. Our results show that the model keeps improving until it hits physical constraints such as camera or motor resolution, or limits on its internal coding capacity. Furthermore, we show that the emerging sensory tuning properties are in line with results on disparity, motion, and motion-in-depth tuning in the visual cortex of mammals. The model suggests that vergence and tracking eye movements can be viewed as fundamentally having the same objective of maximizing the coding efficiency of the visual system and that they can be learned and calibrated jointly through AEC.

摘要

我们提出了一种用于自主同时学习主动双目视觉和运动视觉的模型。该模型基于主动高效编码(AEC)框架,这是经典高效编码理论对主动感知的最新推广。该模型通过稀疏编码学习如何有效地编码由三维空间中移动的物体产生的传入视觉信号。同时,它学习如何产生眼球运动,以进一步提高感官编码的效率。这种学习由最大化系统编码效率的内在动机驱动。我们使用模拟在人形机器人iCub上测试了我们的方法。该模型展示了精确物体注视的自校准和对移动物体的跟踪。我们的结果表明,该模型会持续改进,直到达到物理限制,如相机或电机分辨率,或其内部编码能力的限制。此外,我们表明,新兴的感官调谐特性与哺乳动物视觉皮层中视差、运动和深度运动调谐的结果一致。该模型表明,辐辏和跟踪眼球运动从根本上可以被视为具有相同的目标,即最大化视觉系统的编码效率,并且它们可以通过AEC联合学习和校准。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/0cda2a817c3f/fnbot-13-00049-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/03d1fe34359d/fnbot-13-00049-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/f531da9dc360/fnbot-13-00049-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/727332771bb3/fnbot-13-00049-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/0f137bb12e6f/fnbot-13-00049-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/c121793fae75/fnbot-13-00049-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/01702b755ade/fnbot-13-00049-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/fabcf601e578/fnbot-13-00049-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/b4bbcd526e8e/fnbot-13-00049-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/e8419c17030c/fnbot-13-00049-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/0cda2a817c3f/fnbot-13-00049-g0010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/03d1fe34359d/fnbot-13-00049-g0001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/f531da9dc360/fnbot-13-00049-g0002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/727332771bb3/fnbot-13-00049-g0003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/0f137bb12e6f/fnbot-13-00049-g0004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/c121793fae75/fnbot-13-00049-g0005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/01702b755ade/fnbot-13-00049-g0006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/fabcf601e578/fnbot-13-00049-g0007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/b4bbcd526e8e/fnbot-13-00049-g0008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/e8419c17030c/fnbot-13-00049-g0009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/528b/6646586/0cda2a817c3f/fnbot-13-00049-g0010.jpg

相似文献

1
Autonomous Development of Active Binocular and Motion Vision Through Active Efficient Coding.通过主动高效编码实现主动双目视觉和运动视觉的自主发展。
Front Neurorobot. 2019 Jul 16;13:49. doi: 10.3389/fnbot.2019.00049. eCollection 2019.
2
Robust active binocular vision through intrinsically motivated learning.通过内在动机学习实现稳健的主动双目视觉。
Front Neurorobot. 2013 Nov 7;7:20. doi: 10.3389/fnbot.2013.00020. eCollection 2013.
3
Joint Learning of Binocularly Driven Saccades and Vergence by Active Efficient Coding.通过主动高效编码实现双眼驱动扫视和聚散的联合学习。
Front Neurorobot. 2017 Nov 3;11:58. doi: 10.3389/fnbot.2017.00058. eCollection 2017.
4
The development of active binocular vision under normal and alternate rearing conditions.正常饲养和交替饲养条件下主动双眼视觉的发育
Elife. 2021 Aug 17;10:e56212. doi: 10.7554/eLife.56212.
5
An active-efficient-coding model of optokinetic nystagmus.视动性眼球震颤的主动高效编码模型。
J Vis. 2016 Nov 1;16(14):10. doi: 10.1167/16.14.10.
6
Coding of smooth eye movements in three-dimensional space by frontal cortex.额叶皮层对三维空间中平滑眼球运动的编码。
Nature. 2002 Sep 12;419(6903):157-62. doi: 10.1038/nature00953.
7
A model of how depth facilitates scene-relative object motion perception.一种关于深度如何促进场景相关物体运动感知的模型。
PLoS Comput Biol. 2019 Nov 14;15(11):e1007397. doi: 10.1371/journal.pcbi.1007397. eCollection 2019 Nov.
8
Mice Discriminate Stereoscopic Surfaces Without Fixating in Depth.老鼠无需注视深度即可辨别立体表面。
J Neurosci. 2019 Oct 9;39(41):8024-8037. doi: 10.1523/JNEUROSCI.0895-19.2019. Epub 2019 Aug 28.
9
The dichoptiscope: an instrument for investigating cues to motion in depth.双眼视镜:一种用于研究深度运动线索的仪器。
J Vis. 2013 Dec 2;13(14):1. doi: 10.1167/13.14.1.
10
Mechanisms of static and dynamic stereopsis in foveal cortex of the rhesus monkey.恒河猴中央凹皮层中静态和动态立体视觉的机制。
J Physiol. 1981 Jun;315:469-92. doi: 10.1113/jphysiol.1981.sp013759.

引用本文的文献

1
A spiking neural network for active efficient coding.一种用于主动高效编码的脉冲神经网络。
Front Robot AI. 2025 Jan 15;11:1435197. doi: 10.3389/frobt.2024.1435197. eCollection 2024.
2
Efficient Temporal Coding in the Early Visual System: Existing Evidence and Future Directions.早期视觉系统中的高效时间编码:现有证据与未来方向。
Front Comput Neurosci. 2022 Jul 4;16:929348. doi: 10.3389/fncom.2022.929348. eCollection 2022.
3
The development of active binocular vision under normal and alternate rearing conditions.正常饲养和交替饲养条件下主动双眼视觉的发育

本文引用的文献

1
Emergence of Binocular Disparity Selectivity through Hebbian Learning.通过赫布学习产生双眼视差选择性。
J Neurosci. 2018 Oct 31;38(44):9563-9578. doi: 10.1523/JNEUROSCI.1259-18.2018. Epub 2018 Sep 21.
2
Activity-dependent development of visual receptive fields.视觉感受野的活动依赖性发育
Curr Opin Neurobiol. 2017 Feb;42:136-143. doi: 10.1016/j.conb.2016.12.007. Epub 2017 Jan 11.
3
An active-efficient-coding model of optokinetic nystagmus.视动性眼球震颤的主动高效编码模型。
Elife. 2021 Aug 17;10:e56212. doi: 10.7554/eLife.56212.
4
A Mirror-Based Active Vision System for Underwater Robots: From the Design to Active Object Tracking Application.一种用于水下机器人的基于镜子的主动视觉系统:从设计到主动目标跟踪应用
Front Robot AI. 2021 Jun 21;8:542717. doi: 10.3389/frobt.2021.542717. eCollection 2021.
5
Neurorobots as a Means Toward Neuroethology and Explainable AI.神经机器人作为通往神经行为学和可解释人工智能的一种手段。
Front Neurorobot. 2020 Oct 19;14:570308. doi: 10.3389/fnbot.2020.570308. eCollection 2020.
J Vis. 2016 Nov 1;16(14):10. doi: 10.1167/16.14.10.
4
3D Visual Response Properties of MSTd Emerge from an Efficient, Sparse Population Code.MSTd的3D视觉反应特性源自一种高效、稀疏的群体编码。
J Neurosci. 2016 Aug 10;36(32):8399-415. doi: 10.1523/JNEUROSCI.0396-16.2016.
5
Neuronal activity is not required for the initial formation and maturation of visual selectivity.神经元活动对于视觉选择性的初始形成和成熟并非必需。
Nat Neurosci. 2015 Dec;18(12):1780-8. doi: 10.1038/nn.4155. Epub 2015 Nov 2.
6
Distribution of independent components of binocular natural images.双眼自然图像独立成分的分布
J Vis. 2015;15(13):6. doi: 10.1167/15.13.6.
7
Stereopsis is adaptive for the natural environment.立体视觉适应自然环境。
Sci Adv. 2015 May;1(4). doi: 10.1126/sciadv.1400254.
8
Area MT encodes three-dimensional motion.体感区 MT 编码三维运动。
J Neurosci. 2014 Nov 19;34(47):15522-33. doi: 10.1523/JNEUROSCI.1081-14.2014.
9
Neural representation of motion-in-depth in area MT.MT 区中深度运动的神经表示。
J Neurosci. 2014 Nov 19;34(47):15508-21. doi: 10.1523/JNEUROSCI.1072-14.2014.
10
Emergence of feature-specific connectivity in cortical microcircuits in the absence of visual experience.在没有视觉体验的情况下,皮层微电路中出现特定特征的连接。
J Neurosci. 2014 Jul 16;34(29):9812-6. doi: 10.1523/JNEUROSCI.0875-14.2014.