• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过优化残差生成对抗网络探索语义室内场景合成中的语义标签分解和数据集大小

Exploration of Semantic Label Decomposition and Dataset Size in Semantic Indoor Scenes Synthesis via Optimized Residual Generative Adversarial Networks.

作者信息

Ibrahem Hatem, Salem Ahmed, Kang Hyun-Soo

机构信息

Department of Information and Communication Engineering, School of Electrical and Computer Engineering, Chungbuk National University, Cheongju-si 28644, Korea.

Electrical Engineering Department, Faculty of Engineering, Assiut University, Assiut 71515, Egypt.

出版信息

Sensors (Basel). 2022 Oct 29;22(21):8306. doi: 10.3390/s22218306.

DOI:10.3390/s22218306
PMID:36366007
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9657195/
Abstract

In this paper, we revisit the paired image-to-image translation using the conditional generative adversarial network, the so-called "Pix2Pix", and propose efficient optimization techniques for the architecture and the training method to maximize the architecture's performance to boost the realism of the generated images. We propose a generative adversarial network-based technique to create new artificial indoor scenes using a user-defined semantic segmentation map as an input to define the location, shape, and category of each object in the scene, exactly similar to Pix2Pix. We train different residual connections-based architectures of the generator and discriminator on the NYU depth-v2 dataset and a selected indoor subset from the ADE20K dataset, showing that the proposed models have fewer parameters, less computational complexity, and can generate better quality images than the state of the art methods following the same technique to generate realistic indoor images. We also prove that using extra specific labels and more training samples increases the quality of the generated images; however, the proposed residual connections-based models can learn better from small datasets (i.e., NYU depth-v2) and can improve the realism of the generated images in training on bigger datasets (i.e., ADE20K indoor subset) in comparison to Pix2Pix. The proposed method achieves an LPIPS value of 0.505 and an FID value of 81.067, generating better quality images than that produced by Pix2Pix and other recent paired Image-to-image translation methods and outperforming them in terms of LPIPS and FID.

摘要

在本文中,我们重新审视了使用条件生成对抗网络(即所谓的“Pix2Pix”)进行配对图像到图像的翻译,并针对该架构和训练方法提出了高效的优化技术,以最大化架构性能,提升生成图像的真实感。我们提出了一种基于生成对抗网络的技术,使用用户定义的语义分割图作为输入来创建新的人工室内场景,以定义场景中每个物体的位置、形状和类别,这与Pix2Pix完全类似。我们在NYU深度v2数据集和从ADE20K数据集中选择的室内子集上训练生成器和判别器的不同基于残差连接的架构,结果表明,与采用相同技术生成逼真室内图像的现有方法相比,所提出的模型具有更少的参数、更低的计算复杂度,并且能够生成质量更高的图像。我们还证明,使用额外的特定标签和更多的训练样本可以提高生成图像的质量;然而,与Pix2Pix相比,所提出的基于残差连接的模型能够从较小的数据集(即NYU深度v2)中更好地学习,并且在更大的数据集(即ADE20K室内子集)上训练时能够提高生成图像的真实感。所提出的方法实现了0.505的LPIPS值和81.067的FID值,生成的图像质量优于Pix2Pix和其他最近的配对图像到图像翻译方法所生成的图像,并且在LPIPS和FID方面优于它们。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/1edd0c872c7e/sensors-22-08306-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/0a2154cfbbcc/sensors-22-08306-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/f99cfa5c94f6/sensors-22-08306-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/59a3251bc5f9/sensors-22-08306-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/6c7555132659/sensors-22-08306-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/74da8d3d43bd/sensors-22-08306-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/10c5ecd699af/sensors-22-08306-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/3370435ed245/sensors-22-08306-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/1edd0c872c7e/sensors-22-08306-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/0a2154cfbbcc/sensors-22-08306-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/f99cfa5c94f6/sensors-22-08306-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/59a3251bc5f9/sensors-22-08306-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/6c7555132659/sensors-22-08306-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/74da8d3d43bd/sensors-22-08306-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/10c5ecd699af/sensors-22-08306-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/3370435ed245/sensors-22-08306-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/e401/9657195/1edd0c872c7e/sensors-22-08306-g008.jpg

相似文献

1
Exploration of Semantic Label Decomposition and Dataset Size in Semantic Indoor Scenes Synthesis via Optimized Residual Generative Adversarial Networks.通过优化残差生成对抗网络探索语义室内场景合成中的语义标签分解和数据集大小
Sensors (Basel). 2022 Oct 29;22(21):8306. doi: 10.3390/s22218306.
2
Retinal image synthesis from multiple-landmarks input with generative adversarial networks.基于生成对抗网络的多标记点视网膜图像合成。
Biomed Eng Online. 2019 May 21;18(1):62. doi: 10.1186/s12938-019-0682-x.
3
A pavement crack synthesis method based on conditional generative adversarial networks.一种基于条件生成对抗网络的路面裂缝合成方法。
Math Biosci Eng. 2024 Jan;21(1):903-923. doi: 10.3934/mbe.2024038. Epub 2022 Dec 21.
4
Brain tumor classification for MRI images using dual-discriminator conditional generative adversarial network.基于双鉴别器条件生成对抗网络的 MRI 图像脑肿瘤分类。
Electromagn Biol Med. 2024 Apr 2;43(1-2):81-94. doi: 10.1080/15368378.2024.2321352. Epub 2024 Mar 10.
5
pix2xray: converting RGB images into X-rays using generative adversarial networks.pix2xray:使用生成对抗网络将 RGB 图像转换为 X 射线。
Int J Comput Assist Radiol Surg. 2020 Jun;15(6):973-980. doi: 10.1007/s11548-020-02159-2. Epub 2020 Apr 27.
6
High-Speed Railway Intruding Object Image Generating with Generative Adversarial Networks.基于生成对抗网络的高速铁路入侵物体图像生成
Sensors (Basel). 2019 Jul 11;19(14):3075. doi: 10.3390/s19143075.
7
StackGAN++: Realistic Image Synthesis with Stacked Generative Adversarial Networks.StackGAN++:基于堆叠生成对抗网络的逼真图像合成
IEEE Trans Pattern Anal Mach Intell. 2019 Aug;41(8):1947-1962. doi: 10.1109/TPAMI.2018.2856256. Epub 2018 Jul 16.
8
Pix2pix Conditional Generative Adversarial Networks for Scheimpflug Camera Color-Coded Corneal Tomography Image Generation.基于 Pix2pix 的条件生成对抗网络的 Scheimpflug 彩色角膜断层成像图生成。
Transl Vis Sci Technol. 2021 Jun 1;10(7):21. doi: 10.1167/tvst.10.7.21.
9
Pulmonary nodule segmentation with CT sample synthesis using adversarial networks.基于对抗网络的 CT 样本合成的肺结节分割。
Med Phys. 2019 Mar;46(3):1218-1229. doi: 10.1002/mp.13349. Epub 2019 Jan 31.
10
Conditional generative adversarial network for 3D rigid-body motion correction in MRI.条件生成对抗网络在 MRI 中用于 3D 刚体运动校正。
Magn Reson Med. 2019 Sep;82(3):901-910. doi: 10.1002/mrm.27772. Epub 2019 Apr 22.

引用本文的文献

1
GANMasker: A Two-Stage Generative Adversarial Network for High-Quality Face Mask Removal.GANMasker:一种用于高质量口罩去除的两阶段生成对抗网络。
Sensors (Basel). 2023 Aug 10;23(16):7094. doi: 10.3390/s23167094.