• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

角点和前景面积交并比损失:在边界框回归中更好地定位小目标。

Corner-Point and Foreground-Area IoU Loss: Better Localization of Small Objects in Bounding Box Regression.

机构信息

School of Electrical Engineering and Intelligentization, DongGuan University of Technology, Dongguan 523000, China.

School of Computer Science and Technology, DongGuan University of Technology, Dongguan 523000, China.

出版信息

Sensors (Basel). 2023 May 22;23(10):4961. doi: 10.3390/s23104961.

DOI:10.3390/s23104961
PMID:37430876
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10223589/
Abstract

Bounding box regression is a crucial step in object detection, directly affecting the localization performance of the detected objects. Especially in small object detection, an excellent bounding box regression loss can significantly alleviate the problem of missing small objects. However, there are two major problems with the broad Intersection over Union (IoU) losses, also known as Broad IoU losses (BIoU losses) in bounding box regression: (i) BIoU losses cannot provide more effective fitting information for predicted boxes as they approach the target box, resulting in slow convergence and inaccurate regression results; (ii) most localization loss functions do not fully utilize the spatial information of the target, namely the target's foreground area, during the fitting process. Therefore, this paper proposes the Corner-point and Foreground-area IoU loss (CFIoU loss) function by delving into the potential for bounding box regression losses to overcome these issues. First, we use the normalized corner point distance between the two boxes instead of the normalized center-point distance used in the BIoU losses, which effectively suppresses the problem of BIoU losses degrading to IoU loss when the two boxes are close. Second, we add adaptive target information to the loss function to provide richer target information to optimize the bounding box regression process, especially for small object detection. Finally, we conducted simulation experiments on bounding box regression to validate our hypothesis. At the same time, we conducted quantitative comparisons of the current mainstream BIoU losses and our proposed CFIoU loss on the small object public datasets VisDrone2019 and SODA-D using the latest anchor-based YOLOv5 and anchor-free YOLOv8 object detection algorithms. The experimental results demonstrate that YOLOv5s (+3.12% Recall, +2.73% mAP@0.5, and +1.91% mAP@0.5:0.95) and YOLOv8s (+1.72% Recall and +0.60% mAP@0.5), both incorporating the CFIoU loss, achieved the highest performance improvement on the VisDrone2019 test set. Similarly, YOLOv5s (+6% Recall, +13.08% mAP@0.5, and +14.29% mAP@0.5:0.95) and YOLOv8s (+3.36% Recall, +3.66% mAP@0.5, and +4.05% mAP@0.5:0.95), both incorporating the CFIoU loss, also achieved the highest performance improvement on the SODA-D test set. These results indicate the effectiveness and superiority of the CFIoU loss in small object detection. Additionally, we conducted comparative experiments by fusing the CFIoU loss and the BIoU loss with the SSD algorithm, which is not proficient in small object detection. The experimental results demonstrate that the SSD algorithm incorporating the CFIoU loss achieved the highest improvement in the AP (+5.59%) and AP75 (+5.37%) metrics, indicating that the CFIoU loss can also improve the performance of algorithms that are not proficient in small object detection.

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/b8b6ee4ae0a9/sensors-23-04961-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/99b153852f5a/sensors-23-04961-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/c3adfb4d831f/sensors-23-04961-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/234a9eedffa1/sensors-23-04961-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/9c5b2dff39e6/sensors-23-04961-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/10c6d75d175a/sensors-23-04961-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/ab593cf9298d/sensors-23-04961-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/d69c171b3f2e/sensors-23-04961-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/b8b6ee4ae0a9/sensors-23-04961-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/99b153852f5a/sensors-23-04961-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/c3adfb4d831f/sensors-23-04961-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/234a9eedffa1/sensors-23-04961-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/9c5b2dff39e6/sensors-23-04961-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/10c6d75d175a/sensors-23-04961-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/ab593cf9298d/sensors-23-04961-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/d69c171b3f2e/sensors-23-04961-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c1fa/10223589/b8b6ee4ae0a9/sensors-23-04961-g008.jpg
摘要

边界框回归是目标检测中的关键步骤,直接影响到检测到的物体的定位性能。特别是在小目标检测中,一个优秀的边界框回归损失可以显著缓解小目标的漏检问题。然而,广泛的交并比(IoU)损失有两个主要问题,也称为边界框回归中的宽交并比损失(BIoU 损失):(i)BIoU 损失在预测框接近目标框时,无法为其提供更有效的拟合信息,导致收敛缓慢,回归结果不准确;(ii)大多数定位损失函数在拟合过程中没有充分利用目标的空间信息,即目标的前景区域。因此,本文通过深入研究边界框回归损失的潜力,提出了角点和前景区域 IoU 损失(CFIoU 损失)函数。首先,我们使用两个框之间的归一化角点距离代替 BIoU 损失中使用的归一化中心点距离,这有效地抑制了 BIoU 损失在两个框接近时退化到 IoU 损失的问题。其次,我们在损失函数中添加了自适应的目标信息,为边界框回归过程提供更丰富的目标信息,特别是对于小目标检测。最后,我们在边界框回归上进行了模拟实验,验证了我们的假设。同时,我们在小物体公共数据集 VisDrone2019 和 SODA-D 上使用最新的基于锚点的 YOLOv5 和无锚点的 YOLOv8 目标检测算法,对当前主流的 BIoU 损失和我们提出的 CFIoU 损失进行了定量比较。实验结果表明,YOLOv5s(+3.12%召回率、+2.73%mAP@0.5 和+1.91%mAP@0.5:0.95)和 YOLOv8s(+1.72%召回率和+0.60%mAP@0.5),都包含 CFIoU 损失,在 VisDrone2019 测试集上取得了最高的性能提升。同样,YOLOv5s(+6%召回率、+13.08%mAP@0.5 和+14.29%mAP@0.5:0.95)和 YOLOv8s(+3.36%召回率、+3.66%mAP@0.5 和+4.05%mAP@0.5:0.95),都包含 CFIoU 损失,在 SODA-D 测试集上也取得了最高的性能提升。这些结果表明 CFIoU 损失在小目标检测中的有效性和优越性。此外,我们还通过将 CFIoU 损失和 BIoU 损失与不擅长小目标检测的 SSD 算法进行融合,进行了对比实验。实验结果表明,包含 CFIoU 损失的 SSD 算法在 AP(+5.59%)和 AP75(+5.37%)指标上取得了最高的提升,表明 CFIoU 损失也可以提高不擅长小目标检测的算法的性能。

相似文献

1
Corner-Point and Foreground-Area IoU Loss: Better Localization of Small Objects in Bounding Box Regression.角点和前景面积交并比损失:在边界框回归中更好地定位小目标。
Sensors (Basel). 2023 May 22;23(10):4961. doi: 10.3390/s23104961.
2
Powerful-IoU: More straightforward and faster bounding box regression loss with a nonmonotonic focusing mechanism.强大交并比(Powerful-IoU):一种具有非单调聚焦机制的更直接、更快的边界框回归损失。
Neural Netw. 2024 Feb;170:276-284. doi: 10.1016/j.neunet.2023.11.041. Epub 2023 Nov 22.
3
Object Detection of Flexible Objects with Arbitrary Orientation Based on Rotation-Adaptive YOLOv5.基于旋转自适应 YOLOv5 的任意方向柔性物体目标检测
Sensors (Basel). 2023 May 20;23(10):4925. doi: 10.3390/s23104925.
4
Elevating Detection Performance in Optical Remote Sensing Image Object Detection: A Dual Strategy with Spatially Adaptive Angle-Aware Networks and Edge-Aware Skewed Bounding Box Loss Function.提升光学遥感图像目标检测中的检测性能:一种采用空间自适应角度感知网络和边缘感知倾斜边界框损失函数的双重策略
Sensors (Basel). 2024 Aug 18;24(16):5342. doi: 10.3390/s24165342.
5
SOD-YOLOv8-Enhancing YOLOv8 for Small Object Detection in Aerial Imagery and Traffic Scenes.SOD-YOLOv8——增强YOLOv8以用于航空图像和交通场景中的小目标检测
Sensors (Basel). 2024 Sep 25;24(19):6209. doi: 10.3390/s24196209.
6
Object Detection for UAV Aerial Scenarios Based on Vectorized IOU.基于矢量化 IOU 的无人机空中场景目标检测。
Sensors (Basel). 2023 Mar 13;23(6):3061. doi: 10.3390/s23063061.
7
Confluence: A Robust Non-IoU Alternative to Non-Maxima Suppression in Object Detection.融合:目标检测中一种强大的非交并比(Non-IoU)非极大值抑制替代方法。
IEEE Trans Pattern Anal Mach Intell. 2023 Oct;45(10):11561-11574. doi: 10.1109/TPAMI.2023.3273210. Epub 2023 Sep 5.
8
HP-YOLOv8: High-Precision Small Object Detection Algorithm for Remote Sensing Images.HP-YOLOv8:用于遥感图像的高精度小目标检测算法
Sensors (Basel). 2024 Jul 26;24(15):4858. doi: 10.3390/s24154858.
9
Oriented Vehicle Detection in Aerial Images Based on YOLOv4.基于 YOLOv4 的航空图像中定向车辆检测。
Sensors (Basel). 2022 Nov 1;22(21):8394. doi: 10.3390/s22218394.
10
Dynamic Label Assignment for Object Detection by Combining Predicted IoUs and Anchor IoUs.通过结合预测交并比和锚框交并比进行目标检测的动态标签分配
J Imaging. 2022 Jul 11;8(7):193. doi: 10.3390/jimaging8070193.

引用本文的文献

1
Exploring the impact of hyperparameter and data augmentation in YOLO V10 for accurate bone fracture detection from X-ray images.探索超参数和数据增强对YOLO V10从X射线图像中准确检测骨折的影响。
Sci Rep. 2025 Mar 21;15(1):9828. doi: 10.1038/s41598-025-93505-4.
2
DYNet: A Printed Book Detection Model Using Dual Kernel Neural Networks.DYNet:一种使用双内核神经网络的印刷书籍检测模型。
Sensors (Basel). 2023 Dec 17;23(24):9880. doi: 10.3390/s23249880.
3
Enhanced YOLOv5: An Efficient Road Object Detection Method.增强型YOLOv5:一种高效的道路目标检测方法。

本文引用的文献

1
Towards Large-Scale Small Object Detection: Survey and Benchmarks.迈向大规模小目标检测:综述与基准
IEEE Trans Pattern Anal Mach Intell. 2023 Nov;45(11):13467-13488. doi: 10.1109/TPAMI.2023.3290594. Epub 2023 Oct 3.
Sensors (Basel). 2023 Oct 10;23(20):8355. doi: 10.3390/s23208355.