• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

结构嵌入对抗网络的场景解析。

Structural inference embedded adversarial networks for scene parsing.

机构信息

College of Computer Science and Technology, Harbin Engineering University, Harbin, HeiLongJiang, China.

School of Aeronautics, Northwestern Polytechnical University, Xi'an, ShaanXi, China.

出版信息

PLoS One. 2018 Apr 12;13(4):e0195114. doi: 10.1371/journal.pone.0195114. eCollection 2018.

DOI:10.1371/journal.pone.0195114
PMID:29649294
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC5896926/
Abstract

Explicit structural inference is one key point to improve the accuracy of scene parsing. Meanwhile, adversarial training method is able to reinforce spatial contiguity in output segmentations. To take both advantages of the structural learning and adversarial training simultaneously, we propose a novel deep learning network architecture called Structural Inference Embedded Adversarial Networks (SIEANs) for pixel-wise scene labeling. The generator of our SIEANs, a novel designed scene parsing network, makes full use of convolutional neural networks and long short-term memory networks to learn the global contextual information of objects in four different directions from RGB-(D) images, which is able to describe the (three-dimensional) spatial distributions of objects in a more comprehensive and accurate way. To further improve the performance, we explore the adversarial training method to optimize the generator along with a discriminator, which can not only detect and correct higher-order inconsistencies between the predicted segmentations and corresponding ground truths, but also exploit full advantages of the generator by fine-tuning its parameters so as to obtain higher consistencies. The experimental results demonstrate that our proposed SIEANs is able to achieve a better performance on PASCAL VOC 2012, SIFT FLOW, PASCAL Person-Part, Cityscapes, Stanford Background, NYUDv2, and SUN-RGBD datasets compared to the most of state-of-the-art methods.

摘要

显式结构推断是提高场景解析精度的关键之一。同时,对抗训练方法能够增强输出分割中的空间连续性。为了同时利用结构学习和对抗训练的优势,我们提出了一种名为结构推理嵌入对抗网络(SIEANs)的新型深度学习网络架构,用于像素级场景标注。我们的 SIEANs 的生成器是一种新颖的场景解析网络,充分利用卷积神经网络和长短时记忆网络从 RGB-(D) 图像中学习四个不同方向的物体全局上下文信息,能够更全面、更准确地描述物体的(三维)空间分布。为了进一步提高性能,我们探索了对抗训练方法来优化生成器及其判别器,这不仅可以检测和纠正预测分割与相应地面实况之间的高阶不一致性,还可以通过微调生成器的参数充分利用其优势,以获得更高的一致性。实验结果表明,与最先进的方法相比,我们提出的 SIEANs 在 PASCAL VOC 2012、SIFT FLOW、PASCAL Person-Part、Cityscapes、Stanford Background、NYUDv2 和 SUN-RGBD 数据集上能够实现更好的性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/5ef39bd7c328/pone.0195114.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/cf9bac9dc05c/pone.0195114.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/453a8d6a842a/pone.0195114.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/ee22356d1bc7/pone.0195114.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/b705db50f4bc/pone.0195114.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/13935068aa51/pone.0195114.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/bec114d19ed9/pone.0195114.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/330ba68175fc/pone.0195114.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/8a2142d8180f/pone.0195114.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/5f9d1f70fc0f/pone.0195114.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/0a4074c5f014/pone.0195114.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/25ba4008c1e3/pone.0195114.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/6c8ba44701fb/pone.0195114.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/5ef39bd7c328/pone.0195114.g013.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/cf9bac9dc05c/pone.0195114.g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/453a8d6a842a/pone.0195114.g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/ee22356d1bc7/pone.0195114.g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/b705db50f4bc/pone.0195114.g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/13935068aa51/pone.0195114.g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/bec114d19ed9/pone.0195114.g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/330ba68175fc/pone.0195114.g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/8a2142d8180f/pone.0195114.g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/5f9d1f70fc0f/pone.0195114.g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/0a4074c5f014/pone.0195114.g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/25ba4008c1e3/pone.0195114.g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/6c8ba44701fb/pone.0195114.g012.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/48d2/5896926/5ef39bd7c328/pone.0195114.g013.jpg

相似文献

1
Structural inference embedded adversarial networks for scene parsing.结构嵌入对抗网络的场景解析。
PLoS One. 2018 Apr 12;13(4):e0195114. doi: 10.1371/journal.pone.0195114. eCollection 2018.
2
Generative adversarial networks with decoder-encoder output noises.生成对抗网络与解码器编码器输出噪声。
Neural Netw. 2020 Jul;127:19-28. doi: 10.1016/j.neunet.2020.04.005. Epub 2020 Apr 9.
3
Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition.空间金字塔池化在深度卷积网络中的视觉识别。
IEEE Trans Pattern Anal Mach Intell. 2015 Sep;37(9):1904-16. doi: 10.1109/TPAMI.2015.2389824.
4
A Two-Stream Deep Fusion Framework for High-Resolution Aerial Scene Classification.基于双流深度融合的高分辨率航空场景分类框架。
Comput Intell Neurosci. 2018 Jan 18;2018:8639367. doi: 10.1155/2018/8639367. eCollection 2018.
5
Generative adversarial networks with mixture of t-distributions noise for diverse image generation.基于 t 分布混合噪声的生成对抗网络用于生成多样化的图像。
Neural Netw. 2020 Feb;122:374-381. doi: 10.1016/j.neunet.2019.11.003. Epub 2019 Nov 18.
6
Marginal Space Deep Learning: Efficient Architecture for Volumetric Image Parsing.边缘空间深度学习:用于体绘制图像分割的高效架构。
IEEE Trans Med Imaging. 2016 May;35(5):1217-1228. doi: 10.1109/TMI.2016.2538802. Epub 2016 Mar 7.
7
Unsupervised Domain Adaptation for Facial Expression Recognition Using Generative Adversarial Networks.基于生成对抗网络的无监督领域自适应人脸表情识别。
Comput Intell Neurosci. 2018 Jul 9;2018:7208794. doi: 10.1155/2018/7208794. eCollection 2018.
8
Adversarial symmetric GANs: Bridging adversarial samples and adversarial networks.对抗对称 GANs:连接对抗样本和对抗网络。
Neural Netw. 2021 Jan;133:148-156. doi: 10.1016/j.neunet.2020.10.016. Epub 2020 Nov 6.
9
Deep learning with convolutional neural network in radiology.放射学中基于卷积神经网络的深度学习。
Jpn J Radiol. 2018 Apr;36(4):257-272. doi: 10.1007/s11604-018-0726-3. Epub 2018 Mar 1.
10
Deep Human Parsing with Active Template Regression.深度人体解析与主动模板回归。
IEEE Trans Pattern Anal Mach Intell. 2015 Dec;37(12):2402-14. doi: 10.1109/TPAMI.2015.2408360.

本文引用的文献

1
DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs.DeepLab:基于深度卷积网络、空洞卷积和全连接条件随机场的语义图像分割。
IEEE Trans Pattern Anal Mach Intell. 2018 Apr;40(4):834-848. doi: 10.1109/TPAMI.2017.2699184. Epub 2017 Apr 27.
2
Unsupervised Joint Feature Learning and Encoding for RGB-D Scene Labeling.基于 RGB-D 场景标注的无监督联合特征学习与编码。
IEEE Trans Image Process. 2015 Nov;24(11):4459-73. doi: 10.1109/TIP.2015.2465133. Epub 2015 Aug 11.
3
Learning hierarchical features for scene labeling.
学习用于场景标注的层次特征。
IEEE Trans Pattern Anal Mach Intell. 2013 Aug;35(8):1915-29. doi: 10.1109/TPAMI.2012.231.
4
Long short-term memory.长短期记忆
Neural Comput. 1997 Nov 15;9(8):1735-80. doi: 10.1162/neco.1997.9.8.1735.