• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

统一的 DeepLabV3+ 用于半暗图像语义分割。

Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation.

机构信息

High Performance Cloud Computing Center (HPC3), Department of Computer and Information Sciences, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Malaysia.

Department of Computer Science, Shaheed Zulfiqar Ali Bhutto Institute of Science and Technology, Karachi 75600, Pakistan.

出版信息

Sensors (Basel). 2022 Jul 15;22(14):5312. doi: 10.3390/s22145312.

DOI:10.3390/s22145312
PMID:35890992
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9324997/
Abstract

Semantic segmentation for accurate visual perception is a critical task in computer vision. In principle, the automatic classification of dynamic visual scenes using predefined object classes remains unresolved. The challenging problems of learning deep convolution neural networks, specifically ResNet-based DeepLabV3+ (the most recent version), are threefold. The problems arise due to (1) biased centric exploitations of filter masks, (2) lower representational power of residual networks due to identity shortcuts, and (3) a loss of spatial relationship by using per-pixel primitives. To solve these problems, we present a proficient approach based on DeepLabV3+, along with an added evaluation metric, namely, Unified DeepLabV3+ and S3core, respectively. The presented unified version reduced the effect of biased exploitations via additional dilated convolution layers with customized dilation rates. We further tackled the problem of representational power by introducing non-linear group normalization shortcuts to solve the focused problem of semi-dark images. Meanwhile, to keep track of the spatial relationships in terms of the global and local contexts, geometrically bunched pixel cues were used. We accumulated all the proposed variants of DeepLabV3+ to propose Unified DeepLabV3+ for accurate visual decisions. Finally, the proposed S3core evaluation metric was based on the weighted combination of three different accuracy measures, i.e., the pixel accuracy, IoU (intersection over union), and Mean BFScore, as robust identification criteria. Extensive experimental analysis performed over a CamVid dataset confirmed the applicability of the proposed solution for autonomous vehicles and robotics for outdoor settings. The experimental analysis showed that the proposed Unified DeepLabV3+ outperformed DeepLabV3+ by a margin of 3% in terms of the class-wise pixel accuracy, along with a higher S3core, depicting the effectiveness of the proposed approach.

摘要

语义分割是计算机视觉中一个重要的研究方向,旨在实现对动态视觉场景的精确感知。目前,利用预定义的目标类别来自动分类动态视觉场景仍然是一个具有挑战性的问题。基于深度卷积神经网络的学习,尤其是基于 ResNet 的 DeepLabV3+(最新版本),存在三个问题。这些问题源于:(1)滤波器掩模的中心偏见利用;(2)由于身份捷径,残差网络的表示能力较低;(3)使用像素级原语导致空间关系丢失。为了解决这些问题,我们提出了一种基于 DeepLabV3+的高效方法,并引入了一个附加的评估指标,即统一的 DeepLabV3+和 S3core。所提出的统一版本通过使用自定义膨胀率的额外扩张卷积层减少了偏置利用的影响。我们通过引入非线性组归一化捷径来解决半暗图像的焦点问题,从而进一步解决表示能力的问题。同时,为了跟踪全局和局部上下文的空间关系,使用了几何聚集的像素线索。我们将 DeepLabV3+的所有变体都集成在一起,提出了统一的 DeepLabV3+,以进行准确的视觉决策。最后,所提出的 S3core 评估指标是基于三种不同的精度度量,即像素精度、IoU(交并比)和平均 BFScore 的加权组合,作为鲁棒的识别标准。在 CamVid 数据集上进行的广泛实验分析证实了该解决方案在自主车辆和机器人领域的适用性,用于户外环境。实验分析表明,所提出的统一 DeepLabV3+在类别像素精度方面比 DeepLabV3+高出 3%,同时 S3core 也更高,这表明了所提出方法的有效性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/1abf6a57c36a/sensors-22-05312-g008a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/a8bb0cb348ee/sensors-22-05312-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/ade9fe62b90c/sensors-22-05312-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/54a733c96130/sensors-22-05312-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/e232ebcb662c/sensors-22-05312-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/0dc2fc2a9019/sensors-22-05312-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/7aaafe1fbb7d/sensors-22-05312-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/acf1ae254aa5/sensors-22-05312-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/1abf6a57c36a/sensors-22-05312-g008a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/a8bb0cb348ee/sensors-22-05312-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/ade9fe62b90c/sensors-22-05312-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/54a733c96130/sensors-22-05312-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/e232ebcb662c/sensors-22-05312-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/0dc2fc2a9019/sensors-22-05312-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/7aaafe1fbb7d/sensors-22-05312-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/acf1ae254aa5/sensors-22-05312-g007a.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/6ae8/9324997/1abf6a57c36a/sensors-22-05312-g008a.jpg

相似文献

1
Unified DeepLabV3+ for Semi-Dark Image Semantic Segmentation.统一的 DeepLabV3+ 用于半暗图像语义分割。
Sensors (Basel). 2022 Jul 15;22(14):5312. doi: 10.3390/s22145312.
2
Automated semantic lung segmentation in chest CT images using deep neural network.使用深度神经网络对胸部CT图像进行自动语义肺部分割
Neural Comput Appl. 2023;35(21):15343-15364. doi: 10.1007/s00521-023-08407-1. Epub 2023 Apr 10.
3
ACDSSNet: Atrous Convolution-Based Deep Semantic Segmentation Network for Efficient Detection of Sickle Cell Anemia.ACDSSNet:基于空洞卷积的深度语义分割网络,用于高效检测镰状细胞贫血。
IEEE J Biomed Health Inform. 2024 Oct;28(10):5676-5684. doi: 10.1109/JBHI.2024.3362843. Epub 2024 Oct 3.
4
A comparative study of pre-trained convolutional neural networks for semantic segmentation of breast tumors in ultrasound.用于超声乳腺肿瘤语义分割的预训练卷积神经网络的比较研究
Comput Biol Med. 2020 Nov;126:104036. doi: 10.1016/j.compbiomed.2020.104036. Epub 2020 Oct 8.
5
Multi-task semantic segmentation of CT images for COVID-19 infections using DeepLabV3+ based on dilated residual network.基于空洞残差网络的 DeepLabV3+ 对 COVID-19 感染 CT 图像的多任务语义分割。
Phys Eng Sci Med. 2022 Jun;45(2):443-455. doi: 10.1007/s13246-022-01110-w. Epub 2022 Mar 14.
6
Cross-Image Pixel Contrasting for Semantic Segmentation.用于语义分割的跨图像像素对比
IEEE Trans Pattern Anal Mach Intell. 2024 Aug;46(8):5398-5412. doi: 10.1109/TPAMI.2024.3367952. Epub 2024 Jul 2.
7
Multi-Scale Deep Neural Network Based on Dilated Convolution for Spacecraft Image Segmentation.基于扩张卷积的多尺度深度神经网络在航天器图像分割中的应用。
Sensors (Basel). 2022 Jun 1;22(11):4222. doi: 10.3390/s22114222.
8
Evaluation of Deep Neural Networks for Semantic Segmentation of Prostate in T2W MRI.基于 T2W MRI 的前列腺语义分割的深度神经网络评估。
Sensors (Basel). 2020 Jun 3;20(11):3183. doi: 10.3390/s20113183.
9
Semantic segmentation method for myocardial contrast echocardiogram based on DeepLabV3+ deep learning architecture.基于 DeepLabV3+深度学习架构的心肌对比超声心动图语义分割方法。
Math Biosci Eng. 2023 Jan;20(2):2081-2093. doi: 10.3934/mbe.2023096. Epub 2022 Nov 14.
10
Research on Ground Object Classification Method of High Resolution Remote-Sensing Images Based on Improved DeeplabV3.基于改进型 DeeplabV3 的高分辨率遥感图像地物分类方法研究
Sensors (Basel). 2022 Oct 2;22(19):7477. doi: 10.3390/s22197477.

引用本文的文献

1
The impact of BTK knockdown on lung adenocarcinoma growth and immune response.布鲁顿酪氨酸激酶(BTK)敲低对肺腺癌生长及免疫反应的影响。
Cancer Sci. 2025 Jun;116(6):1550-1564. doi: 10.1111/cas.16394. Epub 2024 Dec 4.
2
FGCN: Image-Fused Point Cloud Semantic Segmentation with Fusion Graph Convolutional Network.FGCN:基于融合图卷积网络的图像融合点云语义分割
Sensors (Basel). 2023 Oct 9;23(19):8338. doi: 10.3390/s23198338.

本文引用的文献

1
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
2
SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.SegNet:一种用于图像分割的深度卷积编解码器架构。
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495. doi: 10.1109/TPAMI.2016.2644615. Epub 2017 Jan 2.
3
Learning hierarchical features for scene labeling.
学习用于场景标注的层次特征。
IEEE Trans Pattern Anal Mach Intell. 2013 Aug;35(8):1915-29. doi: 10.1109/TPAMI.2012.231.