• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MRSNet:基于多分辨率尺度特征融合的通用密度计数网络。

MRSNet: Multi-Resolution Scale Feature Fusion-Based Universal Density Counting Network.

作者信息

Zhang Yi, Song Wei, Shao Mingyue, Liu Xiangchun

机构信息

School of Information and Engineering, Minzu University of China, Beijing 100081, China.

Language Information Security Research Center, Institute of National Security MUC, Minzu University of China, Beijing 100081, China.

出版信息

Sensors (Basel). 2024 Sep 14;24(18):5974. doi: 10.3390/s24185974.

DOI:10.3390/s24185974
PMID:39338718
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11436111/
Abstract

This study focuses on the problem of dense object counting. In dense scenes, variations in object scales and uneven distributions greatly hinder counting accuracy. The current methods, whether CNNs with fixed convolutional kernel sizes or Transformers with fixed attention sizes, struggle to handle such variability effectively. Lower-resolution features are more sensitive to larger objects closer to the camera, while higher-resolution features are more efficient for smaller objects further away. Thus, preserving features that carry the most relevant information at each scale is crucial for improving counting precision. Motivated by this, we propose a multi-resolution scale feature fusion-based universal density counting network (MRSNet). It utilizes independent modules to process high- and low-resolution features, adaptively adjusts receptive field sizes, and incorporates dynamic sparse attention mechanisms to optimize feature information at each resolution, by integrating optimal features across multiple scales into density maps for counting evaluation. Our proposed network effectively mitigates issues caused by large variations in object scales, thereby enhancing counting accuracy. Furthermore, extensive quantitative analyses on six public datasets demonstrate the algorithm's strong generalization ability in handling diverse object scale variations.

摘要

本研究聚焦于密集物体计数问题。在密集场景中,物体尺度的变化和分布不均极大地阻碍了计数精度。当前的方法,无论是具有固定卷积核大小的卷积神经网络(CNNs)还是具有固定注意力大小的Transformer,都难以有效处理这种变化性。较低分辨率的特征对靠近相机的较大物体更敏感,而较高分辨率的特征对较远的较小物体更有效。因此,在每个尺度上保留携带最相关信息的特征对于提高计数精度至关重要。受此启发,我们提出了一种基于多分辨率尺度特征融合的通用密度计数网络(MRSNet)。它利用独立模块处理高分辨率和低分辨率特征,自适应调整感受野大小,并结合动态稀疏注意力机制在每个分辨率下优化特征信息,通过将多个尺度的最优特征整合到密度图中进行计数评估。我们提出的网络有效地减轻了由物体尺度的巨大变化引起的问题,从而提高了计数精度。此外,在六个公共数据集上进行的广泛定量分析证明了该算法在处理各种物体尺度变化方面具有很强的泛化能力。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/7392a7af6fa8/sensors-24-05974-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/af022ec5be0b/sensors-24-05974-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/cbde8b89d02b/sensors-24-05974-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/795d2974996d/sensors-24-05974-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/17dd79311d97/sensors-24-05974-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/ca557a0c6fe9/sensors-24-05974-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/7bf1152f3418/sensors-24-05974-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/ce1cabff8f09/sensors-24-05974-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/7392a7af6fa8/sensors-24-05974-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/af022ec5be0b/sensors-24-05974-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/cbde8b89d02b/sensors-24-05974-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/795d2974996d/sensors-24-05974-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/17dd79311d97/sensors-24-05974-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/ca557a0c6fe9/sensors-24-05974-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/7bf1152f3418/sensors-24-05974-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/ce1cabff8f09/sensors-24-05974-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/3244/11436111/7392a7af6fa8/sensors-24-05974-g008.jpg

相似文献

1
MRSNet: Multi-Resolution Scale Feature Fusion-Based Universal Density Counting Network.MRSNet:基于多分辨率尺度特征融合的通用密度计数网络。
Sensors (Basel). 2024 Sep 14;24(18):5974. doi: 10.3390/s24185974.
2
HADF-Crowd: A Hierarchical Attention-Based Dense Feature Extraction Network for Single-Image Crowd Counting.基于分层注意力的密集特征提取网络用于单幅图像人群计数(HADF-Crowd)。
Sensors (Basel). 2021 May 17;21(10):3483. doi: 10.3390/s21103483.
3
An Adaptive Multi-Scale Network Based on Depth Information for Crowd Counting.一种基于深度信息的自适应多尺度人群计数网络
Sensors (Basel). 2023 Sep 11;23(18):7805. doi: 10.3390/s23187805.
4
Counting dense object of multiple types based on feature enhancement.基于特征增强的多类型密集目标计数
Front Neurorobot. 2024 May 16;18:1383943. doi: 10.3389/fnbot.2024.1383943. eCollection 2024.
5
COMAL: compositional multi-scale feature enhanced learning for crowd counting.COMAL:用于人群计数的组合多尺度特征增强学习
Multimed Tools Appl. 2022;81(15):20541-20560. doi: 10.1007/s11042-022-12249-9. Epub 2022 Mar 11.
6
An effective modular approach for crowd counting in an image using convolutional neural networks.基于卷积神经网络的图像人群计数的有效模块化方法。
Sci Rep. 2022 Apr 6;12(1):5795. doi: 10.1038/s41598-022-09685-w.
7
Crowd Counting Based on Multiscale Spatial Guided Perception Aggregation Network.基于多尺度空间引导感知聚合网络的人群计数
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):17465-17478. doi: 10.1109/TNNLS.2023.3304348. Epub 2024 Dec 2.
8
Embedding Perspective Analysis Into Multi-Column Convolutional Neural Network for Crowd Counting.将视角分析嵌入多列卷积神经网络进行人群计数
IEEE Trans Image Process. 2021;30:1395-1407. doi: 10.1109/TIP.2020.3043122. Epub 2020 Dec 29.
9
Cascaded parallel crowd counting network with multi-resolution collaborative representation.具有多分辨率协作表示的级联并行人群计数网络
Appl Intell (Dordr). 2023;53(3):3002-3016. doi: 10.1007/s10489-022-03639-5. Epub 2022 May 19.
10
Enhanced YOLOv5: An Efficient Road Object Detection Method.增强型YOLOv5:一种高效的道路目标检测方法。
Sensors (Basel). 2023 Oct 10;23(20):8355. doi: 10.3390/s23208355.

本文引用的文献

1
Redesigning Multi-Scale Neural Network for Crowd Counting.重新设计用于人群计数的多尺度神经网络。
IEEE Trans Image Process. 2023;32:3664-3678. doi: 10.1109/TIP.2023.3289290. Epub 2023 Jul 4.
2
Decoupled Two-Stage Crowd Counting and Beyond.解耦的两阶段人群计数及超越。
IEEE Trans Image Process. 2021;30:2862-2875. doi: 10.1109/TIP.2021.3055631. Epub 2021 Feb 12.
3
Locality-Aware Crowd Counting.局部感知人群计数
IEEE Trans Pattern Anal Mach Intell. 2022 Jul;44(7):3602-3613. doi: 10.1109/TPAMI.2021.3056518. Epub 2022 Jun 3.
4
Kernel-Based Density Map Generation for Dense Object Counting.用于密集目标计数的基于核的密度图生成
IEEE Trans Pattern Anal Mach Intell. 2022 Mar;44(3):1357-1370. doi: 10.1109/TPAMI.2020.3022878. Epub 2022 Feb 3.
5
NWPU-Crowd: A Large-Scale Benchmark for Crowd Counting and Localization.西工大人群计数数据集:大规模人群计数和定位基准数据集
IEEE Trans Pattern Anal Mach Intell. 2021 Jun;43(6):2141-2149. doi: 10.1109/TPAMI.2020.3013269. Epub 2021 May 11.
6
Counting Cattle in UAV Images-Dealing with Clustered Animals and Animal/Background Contrast Changes.无人机图像中的牲畜计数——处理聚类动物和动物/背景对比度变化。
Sensors (Basel). 2020 Apr 10;20(7):2126. doi: 10.3390/s20072126.
7
Locate, Size, and Count: Accurately Resolving People in Dense Crowds via Detection.定位、大小和计数:通过检测准确解析密集人群中的人员。
IEEE Trans Pattern Anal Mach Intell. 2021 Aug;43(8):2739-2751. doi: 10.1109/TPAMI.2020.2974830. Epub 2021 Jul 1.
8
TasselNetv2: in-field counting of wheat spikes with context-augmented local regression networks.TasselNetv2:使用上下文增强局部回归网络对小麦穗进行田间计数。
Plant Methods. 2019 Dec 11;15:150. doi: 10.1186/s13007-019-0537-2. eCollection 2019.
9
TasselNet: counting maize tassels in the wild via local counts regression network.TasselNet:通过局部计数回归网络对野外玉米雄穗进行计数
Plant Methods. 2017 Nov 1;13:79. doi: 10.1186/s13007-017-0224-0. eCollection 2017.
10
Object detection with discriminatively trained part-based models.基于判别式训练的部件模型的目标检测。
IEEE Trans Pattern Anal Mach Intell. 2010 Sep;32(9):1627-45. doi: 10.1109/TPAMI.2009.167.