• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于属性分解生成对抗网络的可控图像合成

Controllable Image Synthesis With Attribute-Decomposed GAN.

作者信息

Pu Guo, Men Yifang, Mao Yiming, Jiang Yuning, Ma Wei-Ying, Lian Zhouhui

出版信息

IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1514-1532. doi: 10.1109/TPAMI.2022.3161985. Epub 2023 Jan 6.

DOI:10.1109/TPAMI.2022.3161985
PMID:35324435
Abstract

This paper proposes Attribute-Decomposed GAN (ADGAN) and its enhanced version (ADGAN++) for controllable image synthesis, which can produce realistic images with desired attributes provided in various source inputs. The core ideas of the proposed ADGAN and ADGAN++ are both to embed component attributes into the latent space as independent codes and thus achieve flexible and continuous control of attributes via mixing and interpolation operations in explicit style representations. The major difference between them is that ADGAN processes all component attributes simultaneously while ADGAN++ utilizes a serial encoding strategy. More specifically, ADGAN consists of two encoding pathways with style block connections and is capable of decomposing the original hard mapping into multiple more accessible subtasks. In the source pathway, component layouts are extracted via a semantic parser and the segmented components are fed into a shared global texture encoder to obtain decomposed latent codes. This strategy allows for the synthesis of more realistic output images and the automatic separation of un-annotated component attributes. Although the original ADGAN works in a delicate and efficient manner, intrinsically it fails to handle the semantic image synthesizing task when the number of attribute categories is huge. To address this problem, ADGAN++ employs the serial encoding of different component attributes to synthesize each part of the target real-world image, and adopts several residual blocks with segmentation guided instance normalization to assemble the synthesized component images and refine the original synthesis result. The two-stage ADGAN++ is designed to alleviate the massive computational costs required when synthesizing real-world images with numerous attributes while maintaining the disentanglement of different attributes to enable flexible control of arbitrary component attributes of the synthesized images. Experimental results demonstrate the proposed methods' superiority over the state of the art in pose transfer, face style transfer, and semantic image synthesis, as well as their effectiveness in the task of component attribute transfer. Our code and data are publicly available at https://github.com/menyifang/ADGAN.

摘要

本文提出了用于可控图像合成的属性分解生成对抗网络(ADGAN)及其增强版本(ADGAN++),它可以根据各种源输入中提供的所需属性生成逼真的图像。所提出的ADGAN和ADGAN++的核心思想都是将组件属性作为独立代码嵌入到潜在空间中,从而通过显式风格表示中的混合和插值操作实现对属性的灵活且连续的控制。它们之间的主要区别在于,ADGAN同时处理所有组件属性,而ADGAN++采用串行编码策略。更具体地说,ADGAN由两个带有风格块连接的编码路径组成,能够将原始的硬映射分解为多个更易于处理的子任务。在源路径中,通过语义解析器提取组件布局,并将分割后的组件输入到共享的全局纹理编码器中以获得分解后的潜在代码。这种策略允许合成更逼真的输出图像,并自动分离未标注的组件属性。尽管原始的ADGAN工作方式精细且高效,但本质上当属性类别数量巨大时,它无法处理语义图像合成任务。为了解决这个问题,ADGAN++采用不同组件属性的串行编码来合成目标真实世界图像的每个部分,并采用几个带有分割引导实例归一化的残差块来组装合成的组件图像并优化原始合成结果。两阶段的ADGAN++旨在减轻在合成具有大量属性的真实世界图像时所需的大量计算成本,同时保持不同属性的解缠结,以实现对合成图像的任意组件属性的灵活控制。实验结果表明,所提出的方法在姿态转移、面部风格转移和语义图像合成方面优于现有技术,并且在组件属性转移任务中也很有效。我们的代码和数据可在https://github.com/menyifang/ADGAN上公开获取。

相似文献

1
Controllable Image Synthesis With Attribute-Decomposed GAN.基于属性分解生成对抗网络的可控图像合成
IEEE Trans Pattern Anal Mach Intell. 2023 Feb;45(2):1514-1532. doi: 10.1109/TPAMI.2022.3161985. Epub 2023 Jan 6.
2
Towards Disentangling Latent Space for Unsupervised Semantic Face Editing.面向无监督语义人脸编辑的潜在空间解缠。
IEEE Trans Image Process. 2022;31:1475-1489. doi: 10.1109/TIP.2022.3142527. Epub 2022 Jan 27.
3
On Symbiosis of Attribute Prediction and Semantic Segmentation.属性预测与语义分割的共生。
IEEE Trans Pattern Anal Mach Intell. 2021 May;43(5):1620-1635. doi: 10.1109/TPAMI.2019.2956039. Epub 2021 Apr 1.
4
Research on DUAL-ADGAN Model for Anomaly Detection Method in Time-Series Data.研究时间序列数据异常检测方法的 DUAL-ADGAN 模型。
Comput Intell Neurosci. 2022 Oct 26;2022:8753323. doi: 10.1155/2022/8753323. eCollection 2022.
5
Image-to-Image Translation With Disentangled Latent Vectors for Face Editing.用于面部编辑的具有解缠潜向量的图像到图像翻译
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14777-14788. doi: 10.1109/TPAMI.2023.3308102. Epub 2023 Nov 3.
6
IAS-NET: Joint intraclassly adaptive GAN and segmentation network for unsupervised cross-domain in neonatal brain MRI segmentation.IAS-NET:用于新生儿脑 MRI 分割的无监督跨领域的联合类内自适应 GAN 和分割网络。
Med Phys. 2021 Nov;48(11):6962-6975. doi: 10.1002/mp.15212. Epub 2021 Sep 25.
7
AttGAN: Facial Attribute Editing by Only Changing What You Want.AttGAN:仅通过改变你想要改变的内容来进行面部属性编辑。
IEEE Trans Image Process. 2019 Nov;28(11):5464-5478. doi: 10.1109/TIP.2019.2916751. Epub 2019 May 20.
8
Coupled Attribute Learning for Heterogeneous Face Recognition.用于异构人脸识别的耦合属性学习
IEEE Trans Neural Netw Learn Syst. 2020 Nov;31(11):4699-4712. doi: 10.1109/TNNLS.2019.2957285. Epub 2020 Oct 29.
9
ShaTure: Shape and Texture Deformation for Human Pose and Attribute Transfer.沙图尔:人体姿态和属性迁移的形状和纹理变形。
IEEE Trans Image Process. 2022;31:2541-2556. doi: 10.1109/TIP.2022.3157146. Epub 2022 Mar 21.
10
MaskFaceGAN: High-Resolution Face Editing With Masked GAN Latent Code Optimization.MaskFaceGAN:基于掩码生成对抗网络潜在代码优化的高分辨率面部编辑
IEEE Trans Image Process. 2023;32:5893-5908. doi: 10.1109/TIP.2023.3326675. Epub 2023 Nov 1.

引用本文的文献

1
Virtual group photography based on scene and person perception.基于场景和人物感知的虚拟群组摄影。
Heliyon. 2023 Dec 13;10(1):e23568. doi: 10.1016/j.heliyon.2023.e23568. eCollection 2024 Jan 15.
2
Deep learning-based optic disc classification is affected by optic-disc tilt.基于深度学习的视盘分类受视盘倾斜的影响。
Sci Rep. 2024 Jan 4;14(1):498. doi: 10.1038/s41598-023-50256-4.
3
A Review of Performance Prediction Based on Machine Learning in Materials Science.基于机器学习的材料科学性能预测综述
Nanomaterials (Basel). 2022 Aug 26;12(17):2957. doi: 10.3390/nano12172957.