• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

适用于增强现实的语义分割新方法。

Novel Method of Semantic Segmentation Applicable to Augmented Reality.

机构信息

Department of Electronic Engineering, Hanbat National University, Daejeon 34158, Korea.

Department of Electronics & Control Engineering, Hanbat National University, Daejeon 34158, Korea.

出版信息

Sensors (Basel). 2020 Mar 20;20(6):1737. doi: 10.3390/s20061737.

DOI:10.3390/s20061737
PMID:32245002
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7146136/
Abstract

This paper proposes a novel method of semantic segmentation, consisting of modified dilated residual network, atrous pyramid pooling module, and backpropagation, that is applicable to augmented reality (AR). In the proposed method, the modified dilated residual network extracts a feature map from the original images and maintains spatial information. The atrous pyramid pooling module places convolutions in parallel and layers feature maps in a pyramid shape to extract objects occupying small areas in the image; these are converted into one channel using a 1 × 1 convolution. Backpropagation compares the semantic segmentation obtained through convolution from the final feature map with the ground truth provided by a database. Losses can be reduced by applying backpropagation to the modified dilated residual network to change the weighting. The proposed method was compared with other methods on the Cityscapes and PASCAL VOC 2012 databases. The proposed method achieved accuracies of 82.8 and 89.8 mean intersection over union (mIOU) and frame rates of 61 and 64.3 frames per second (fps) for the Cityscapes and PASCAL VOC 2012 databases, respectively. These results prove the applicability of the proposed method for implementing natural AR applications at actual speeds because the frame rate is greater than 60 fps.

摘要

本文提出了一种新的语义分割方法,该方法由改进的空洞残差网络、空洞金字塔池化模块和反向传播组成,适用于增强现实(AR)。在提出的方法中,改进的空洞残差网络从原始图像中提取特征图并保持空间信息。空洞金字塔池化模块并行放置卷积并将特征图分层为金字塔形状,以提取图像中占据小区域的对象;这些对象使用 1×1 卷积转换为一个通道。反向传播将从最终特征图卷积获得的语义分割与数据库提供的地面真值进行比较。通过将反向传播应用于改进的空洞残差网络来改变权重,可以减少损失。在 Cityscapes 和 PASCAL VOC 2012 数据库上,将提出的方法与其他方法进行了比较。对于 Cityscapes 和 PASCAL VOC 2012 数据库,提出的方法的平均交并比(mIOU)分别为 82.8 和 89.8,帧率分别为 61 和 64.3 帧/秒(fps)。这些结果证明了所提出的方法在实际速度下实现自然 AR 应用的适用性,因为帧率大于 60 fps。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/555fdb958e15/sensors-20-01737-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/53f398e1e7ce/sensors-20-01737-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/c82c0a891209/sensors-20-01737-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/120b0f216184/sensors-20-01737-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/3e143e510d03/sensors-20-01737-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/bf3fc317fbdb/sensors-20-01737-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/85c3a349913a/sensors-20-01737-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/14d13a520870/sensors-20-01737-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/66b8ab48331b/sensors-20-01737-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/0ee18519f884/sensors-20-01737-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/317d6ced81d8/sensors-20-01737-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/4f6779f721a7/sensors-20-01737-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/c4f742c04a8a/sensors-20-01737-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/3afc7f747ce2/sensors-20-01737-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/542b8cb697ba/sensors-20-01737-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/6efa21d34358/sensors-20-01737-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/5490f19bae04/sensors-20-01737-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/b20c2b536ce5/sensors-20-01737-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/4310efde0c92/sensors-20-01737-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/46cfaff73ce4/sensors-20-01737-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/555fdb958e15/sensors-20-01737-g020.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/53f398e1e7ce/sensors-20-01737-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/c82c0a891209/sensors-20-01737-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/120b0f216184/sensors-20-01737-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/3e143e510d03/sensors-20-01737-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/bf3fc317fbdb/sensors-20-01737-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/85c3a349913a/sensors-20-01737-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/14d13a520870/sensors-20-01737-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/66b8ab48331b/sensors-20-01737-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/0ee18519f884/sensors-20-01737-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/317d6ced81d8/sensors-20-01737-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/4f6779f721a7/sensors-20-01737-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/c4f742c04a8a/sensors-20-01737-g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/3afc7f747ce2/sensors-20-01737-g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/542b8cb697ba/sensors-20-01737-g014.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/6efa21d34358/sensors-20-01737-g015.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/5490f19bae04/sensors-20-01737-g016.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/b20c2b536ce5/sensors-20-01737-g017.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/4310efde0c92/sensors-20-01737-g018.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/46cfaff73ce4/sensors-20-01737-g019.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e136/7146136/555fdb958e15/sensors-20-01737-g020.jpg

相似文献

1
Novel Method of Semantic Segmentation Applicable to Augmented Reality.适用于增强现实的语义分割新方法。
Sensors (Basel). 2020 Mar 20;20(6):1737. doi: 10.3390/s20061737.
2
A multiple-channel and atrous convolution network for ultrasound image segmentation.一种用于超声图像分割的多通道多孔卷积网络。
Med Phys. 2020 Dec;47(12):6270-6285. doi: 10.1002/mp.14512. Epub 2020 Oct 18.
3
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
4
Semantic Segmentation Network Based on Adaptive Attention and Deep Fusion Utilizing a Multi-Scale Dilated Convolutional Pyramid.基于自适应注意力和深度融合的语义分割网络:利用多尺度扩张卷积金字塔
Sensors (Basel). 2024 Aug 16;24(16):5305. doi: 10.3390/s24165305.
5
[Segmentation of ground glass pulmonary nodules using full convolution residual network based on atrous spatial pyramid pooling structure and attention mechanism].基于空洞空间金字塔池化结构和注意力机制的全卷积残差网络用于磨玻璃肺结节分割
Sheng Wu Yi Xue Gong Cheng Xue Za Zhi. 2022 Jun 25;39(3):441-451. doi: 10.7507/1001-5515.202010051.
6
Waterfall Atrous Spatial Pooling Architecture for Efficient Semantic Segmentation.用于高效语义分割的瀑布空洞空间池化架构。
Sensors (Basel). 2019 Dec 5;19(24):5361. doi: 10.3390/s19245361.
7
ADR-Net: Context extraction network based on M-Net for medical image segmentation.ADR-Net:基于M-Net的医学图像分割上下文提取网络。
Med Phys. 2020 Sep;47(9):4254-4264. doi: 10.1002/mp.14364. Epub 2020 Aug 2.
8
Multi-Scale Deep Neural Network Based on Dilated Convolution for Spacecraft Image Segmentation.基于扩张卷积的多尺度深度神经网络在航天器图像分割中的应用。
Sensors (Basel). 2022 Jun 1;22(11):4222. doi: 10.3390/s22114222.
9
High-Resolution Aerial Imagery Semantic Labeling with Dense Pyramid Network.高分辨率航空影像语义标注的密集金字塔网络方法
Sensors (Basel). 2018 Nov 5;18(11):3774. doi: 10.3390/s18113774.
10
A fish image segmentation methodology in aquaculture environment based on multi-feature fusion model.基于多特征融合模型的水产养殖环境下的鱼类图像分割方法。
Mar Environ Res. 2023 Sep;190:106085. doi: 10.1016/j.marenvres.2023.106085. Epub 2023 Jul 7.

引用本文的文献

1
ThelR547v1-An Asymmetric Dilated Convolutional Neural Network for Real-time Semantic Segmentation of Horticultural Crops.ThelR547v1-一种用于园艺作物实时语义分割的非对称扩张卷积神经网络。
Sensors (Basel). 2022 Nov 15;22(22):8807. doi: 10.3390/s22228807.
2
COVID-19 Isolation Control Proposal UAV and UGV for Crowded Indoor Environments: Assistive Robots in the Shopping Malls.COVID-19 隔离控制方案 用于拥挤室内环境的无人机和地面机器人:购物中心的辅助机器人。
Front Public Health. 2022 May 31;10:855994. doi: 10.3389/fpubh.2022.855994. eCollection 2022.
3
Automated Knee MR Images Segmentation of Anterior Cruciate Ligament Tears.

本文引用的文献

1
Stacked Deconvolutional Network for Semantic Segmentation.用于语义分割的堆叠去卷积网络
IEEE Trans Image Process. 2019 Jan 25. doi: 10.1109/TIP.2019.2895460.
2
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
3
SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.
自动分割膝关节前交叉韧带撕裂的磁共振图像。
Sensors (Basel). 2022 Feb 17;22(4):1552. doi: 10.3390/s22041552.
4
Special Issue on "Smart City and Smart Infrastructure".特刊:“智慧城市与智能基础设施”
Sensors (Basel). 2021 Oct 25;21(21):7064. doi: 10.3390/s21217064.
5
Pancreatic Cancer Survival Prediction: A Survey of the State-of-the-Art.胰腺癌生存预测:现状调查。
Comput Math Methods Med. 2021 Sep 30;2021:1188414. doi: 10.1155/2021/1188414. eCollection 2021.
6
Knowledge Distillation for Semantic Segmentation Using Channel and Spatial Correlations and Adaptive Cross Entropy.基于通道与空间相关性及自适应交叉熵的语义分割知识蒸馏
Sensors (Basel). 2020 Aug 17;20(16):4616. doi: 10.3390/s20164616.
SegNet:一种用于图像分割的深度卷积编解码器架构。
IEEE Trans Pattern Anal Mach Intell. 2017 Dec;39(12):2481-2495. doi: 10.1109/TPAMI.2016.2644615. Epub 2017 Jan 2.