• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于合并语义属性和外观特征的多目标跟踪脑策略算法。

Brain Strategy Algorithm for Multiple Object Tracking Based on Merging Semantic Attributes and Appearance Features.

机构信息

Faculty of Computer & Artificial Intelligence, Benha University, Benha 13511, Egypt.

Intoolab Ltd., London WC2H 9JQ, UK.

出版信息

Sensors (Basel). 2021 Nov 16;21(22):7604. doi: 10.3390/s21227604.

DOI:10.3390/s21227604
PMID:34833680
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8625767/
Abstract

The human brain can effortlessly perform vision processes using the visual system, which helps solve multi-object tracking (MOT) problems. However, few algorithms simulate human strategies for solving MOT. Therefore, devising a method that simulates human activity in vision has become a good choice for improving MOT results, especially occlusion. Eight brain strategies have been studied from a cognitive perspective and imitated to build a novel algorithm. Two of these strategies gave our algorithm novel and outstanding results, rescuing saccades and stimulus attributes. First, rescue saccades were imitated by detecting the occlusion state in each frame, representing the critical situation that the human brain saccades toward. Then, stimulus attributes were mimicked by using semantic attributes to reidentify the person in these occlusion states. Our algorithm favourably performs on the MOT17 dataset compared to state-of-the-art trackers. In addition, we created a new dataset of 40,000 images, 190,000 annotations and 4 classes to train the detection model to detect occlusion and semantic attributes. The experimental results demonstrate that our new dataset achieves an outstanding performance on the scaled YOLOv4 detection model by achieving a 0.89 mAP 0.5.

摘要

人类大脑可以轻松地利用视觉系统进行视觉处理,这有助于解决多目标跟踪 (MOT) 问题。然而,很少有算法模拟人类解决 MOT 的策略。因此,设计一种模拟人类视觉活动的方法已成为提高 MOT 结果的一个很好的选择,尤其是在解决遮挡问题方面。本文从认知的角度研究了 8 种大脑策略,并对其进行模仿,以构建一种新的算法。其中两种策略为我们的算法带来了新颖和出色的结果,即挽救扫视和刺激属性。首先,通过检测每一帧中的遮挡状态来模仿挽救扫视,以表示大脑向关键情况扫视的状态。然后,通过使用语义属性来重新识别这些遮挡状态中的人,模仿刺激属性。与最先进的跟踪器相比,我们的算法在 MOT17 数据集上表现良好。此外,我们创建了一个包含 40000 张图像、190000 个标注和 4 个类别的新数据集,以训练检测模型来检测遮挡和语义属性。实验结果表明,我们的新数据集在缩放的 YOLOv4 检测模型上取得了卓越的性能,达到了 0.89 的 mAP0.5。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/616885f5944b/sensors-21-07604-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/9c01d24d5a08/sensors-21-07604-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/9608e73208b3/sensors-21-07604-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/0f72120dacfd/sensors-21-07604-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/ad2da7858b0e/sensors-21-07604-g0A4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/e9e6f1782c18/sensors-21-07604-g0A5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/237c38984951/sensors-21-07604-g0A6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/b529d9d4633c/sensors-21-07604-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/fc45eb5353fc/sensors-21-07604-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/6cb4d255701d/sensors-21-07604-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/1f0c02a45f82/sensors-21-07604-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/ead1f10be3c1/sensors-21-07604-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/da7f153e80ae/sensors-21-07604-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/86ac040d6d3a/sensors-21-07604-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/48248ce0e302/sensors-21-07604-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/f2a249e02df3/sensors-21-07604-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/f656921169af/sensors-21-07604-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/6feec0d54116/sensors-21-07604-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/593ab42f80a9/sensors-21-07604-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/616885f5944b/sensors-21-07604-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/9c01d24d5a08/sensors-21-07604-g0A1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/9608e73208b3/sensors-21-07604-g0A2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/0f72120dacfd/sensors-21-07604-g0A3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/ad2da7858b0e/sensors-21-07604-g0A4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/e9e6f1782c18/sensors-21-07604-g0A5.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/237c38984951/sensors-21-07604-g0A6.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/b529d9d4633c/sensors-21-07604-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/fc45eb5353fc/sensors-21-07604-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/6cb4d255701d/sensors-21-07604-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/1f0c02a45f82/sensors-21-07604-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/ead1f10be3c1/sensors-21-07604-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/da7f153e80ae/sensors-21-07604-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/86ac040d6d3a/sensors-21-07604-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/48248ce0e302/sensors-21-07604-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/f2a249e02df3/sensors-21-07604-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/f656921169af/sensors-21-07604-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/6feec0d54116/sensors-21-07604-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/593ab42f80a9/sensors-21-07604-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/7cb9/8625767/616885f5944b/sensors-21-07604-g013.jpg

相似文献

1
Brain Strategy Algorithm for Multiple Object Tracking Based on Merging Semantic Attributes and Appearance Features.基于合并语义属性和外观特征的多目标跟踪脑策略算法。
Sensors (Basel). 2021 Nov 16;21(22):7604. doi: 10.3390/s21227604.
2
AttMOT: Improving Multiple-Object Tracking by Introducing Auxiliary Pedestrian Attributes.AttMOT:通过引入辅助行人属性改进多目标跟踪
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):5454-5468. doi: 10.1109/TNNLS.2024.3384446. Epub 2025 Feb 28.
3
Multi-Target Tracking Based on a Combined Attention Mechanism and Occlusion Sensing in a Behavior-Analysis System.基于行为分析系统中的联合注意力机制和遮挡感知的多目标跟踪。
Sensors (Basel). 2023 Mar 8;23(6):2956. doi: 10.3390/s23062956.
4
Spatial-Semantic and Temporal Attention Mechanism-Based Online Multi-Object Tracking.基于空间-语义和时间注意机制的在线多目标跟踪。
Sensors (Basel). 2020 Mar 16;20(6):1653. doi: 10.3390/s20061653.
5
Joint Semantic and Latent Attribute Modelling for Cross-Class Transfer Learning.跨类迁移学习的联合语义和潜在属性建模。
IEEE Trans Pattern Anal Mach Intell. 2018 Jul;40(7):1625-1638. doi: 10.1109/TPAMI.2017.2723882. Epub 2017 Jul 6.
6
Pixel-Guided Association for Multi-Object Tracking.用于多目标跟踪的像素引导关联
Sensors (Basel). 2022 Nov 18;22(22):8922. doi: 10.3390/s22228922.
7
Tracking Beyond Detection: Learning a Global Response Map for End-to-End Multi-Object Tracking.超越检测的跟踪:学习用于端到端多目标跟踪的全局响应图
IEEE Trans Image Process. 2021;30:8222-8235. doi: 10.1109/TIP.2021.3113169. Epub 2021 Sep 30.
8
STMMOT: Advancing multi-object tracking through spatiotemporal memory networks and multi-scale attention pyramids.STMMOT:通过时空记忆网络和多尺度注意力金字塔推进多目标跟踪
Neural Netw. 2023 Nov;168:363-379. doi: 10.1016/j.neunet.2023.09.047. Epub 2023 Sep 29.
9
Coupled Attribute Learning for Heterogeneous Face Recognition.用于异构人脸识别的耦合属性学习
IEEE Trans Neural Netw Learn Syst. 2020 Nov;31(11):4699-4712. doi: 10.1109/TNNLS.2019.2957285. Epub 2020 Oct 29.
10
An anti-occlusion optimization algorithm for multiple pedestrian tracking.一种用于多人行跟踪的防堵塞优化算法。
PLoS One. 2024 Jan 31;19(1):e0291538. doi: 10.1371/journal.pone.0291538. eCollection 2024.

引用本文的文献

1
Improved DeepSORT-Based Object Tracking in Foggy Weather for AVs Using Sematic Labels and Fused Appearance Feature Network.基于改进深度排序的自动驾驶车辆在雾天中使用语义标签和融合外观特征网络的目标跟踪
Sensors (Basel). 2024 Jul 19;24(14):4692. doi: 10.3390/s24144692.
2
Experimental Studies of the Attention Processing Model in Multiple Object Tracking Task.多目标跟踪任务中注意加工模型的实验研究
Brain Sci. 2022 Dec 8;12(12):1686. doi: 10.3390/brainsci12121686.
3
Special Issue "Computer Aided Diagnosis Sensors".特刊:计算机辅助诊断传感器

本文引用的文献

1
Multiple-target tracking in human and machine vision.人类和机器视觉中的多目标跟踪。
PLoS Comput Biol. 2020 Apr 9;16(4):e1007698. doi: 10.1371/journal.pcbi.1007698. eCollection 2020 Apr.
2
Data Association for Multi-Object Tracking via Deep Neural Networks.基于深度神经网络的多目标跟踪数据关联。
Sensors (Basel). 2019 Jan 29;19(3):559. doi: 10.3390/s19030559.
3
Temporal dynamics of access to amodal representations of category-level conceptual information.类别级概念信息的非模态表示的访问的时间动态。
Sensors (Basel). 2022 Oct 21;22(20):8052. doi: 10.3390/s22208052.
Sci Rep. 2019 Jan 18;9(1):239. doi: 10.1038/s41598-018-37429-2.
4
Confidence-Based Data Association and Discriminative Deep Appearance Learning for Robust Online Multi-Object Tracking.基于置信度的数据关联和判别式深度表观学习的鲁棒在线多目标跟踪。
IEEE Trans Pattern Anal Mach Intell. 2018 Mar;40(3):595-610. doi: 10.1109/TPAMI.2017.2691769. Epub 2017 Apr 6.
5
Semantic attributes are encoded in human electrocorticographic signals during visual object recognition.语义属性在人类视觉物体识别的脑电信号中被编码。
Neuroimage. 2017 Mar 1;148:318-329. doi: 10.1016/j.neuroimage.2016.12.074. Epub 2017 Jan 11.
6
A Biologically Inspired Appearance Model for Robust Visual Tracking.基于生物启发的外观模型的鲁棒视觉跟踪。
IEEE Trans Neural Netw Learn Syst. 2017 Oct;28(10):2357-2370. doi: 10.1109/TNNLS.2016.2586194. Epub 2016 Jul 19.
7
Robust Visual Tracking via Convolutional Networks Without Training.基于卷积网络的无需训练的鲁棒视觉跟踪。
IEEE Trans Image Process. 2016 Apr;25(4):1779-92. doi: 10.1109/TIP.2016.2531283. Epub 2016 Feb 18.
8
Brain Activation of Identity Switching in Multiple Identity Tracking Task.多重身份追踪任务中身份切换的大脑激活情况
PLoS One. 2015 Dec 23;10(12):e0145489. doi: 10.1371/journal.pone.0145489. eCollection 2015.
9
People recognition: a historical/anthropological perspective.人物识别:历史/人类学视角。
Behav Neurol. 1993;6(2):99-105. doi: 10.3233/BEN-1993-6205.
10
Memory-based multiagent coevolution modeling for robust moving object tracking.用于稳健移动目标跟踪的基于记忆的多智能体协同进化建模
ScientificWorldJournal. 2013 Jun 16;2013:793013. doi: 10.1155/2013/793013. Print 2013.