• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

FEditNet++:基于相关属性解缠的生成对抗网络空间中潜在语义的少样本编辑

FEditNet++: Few-Shot Editing of Latent Semantics in GAN Spaces With Correlated Attribute Disentanglement.

作者信息

Yi Ran, Hu Teng, Xia Mengfei, Tang Yizhe, Liu Yong-Jin

出版信息

IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):9975-9990. doi: 10.1109/TPAMI.2024.3432529. Epub 2024 Nov 6.

DOI:10.1109/TPAMI.2024.3432529
PMID:39042535
Abstract

Generative Adversarial Networks have achieved significant advancements in generating and editing high-resolution images. However, most methods suffer from either requiring extensive labeled datasets or strong prior knowledge. It is also challenging for them to disentangle correlated attributes with few-shot data. In this paper, we propose FEditNet++, a GAN-based approach to explore latent semantics. It aims to enable attribute editing with limited labeled data and disentangle the correlated attributes. We propose a layer-wise feature contrastive objective, which takes into consideration content consistency and facilitates the invariance of the unrelated attributes before and after editing. Furthermore, we harness the knowledge from the pretrained discriminative model to prevent overfitting. In particular, to solve the entanglement problem between the correlated attributes from data and semantic latent correlation, we extend our model to jointly optimize multiple attributes and propose a novel decoupling loss and cross-assessment loss to disentangle them from both latent and image space. We further propose a novel-attribute disentanglement strategy to enable editing of novel attributes with unknown entanglements. Finally, we extend our model to accurately edit the fine-grained attributes. Qualitative and quantitative assessments demonstrate that our method outperforms state-of-the-art approaches across various datasets, including CelebA-HQ, RaFD, Danbooru2018 and LSUN Church.

摘要

生成对抗网络在生成和编辑高分辨率图像方面取得了重大进展。然而,大多数方法要么需要大量的标记数据集,要么需要强大的先验知识。对于它们来说,利用少样本数据解开相关属性也具有挑战性。在本文中,我们提出了FEditNet++,一种基于GAN的探索潜在语义的方法。它旨在使用有限的标记数据进行属性编辑,并解开相关属性。我们提出了一种逐层特征对比目标,该目标考虑了内容一致性,并促进了编辑前后无关属性的不变性。此外,我们利用预训练判别模型的知识来防止过拟合。特别是,为了解决数据中相关属性与语义潜在相关性之间的纠缠问题,我们扩展了模型以联合优化多个属性,并提出了一种新颖的解耦损失和交叉评估损失,以从潜在空间和图像空间中解开它们。我们进一步提出了一种新颖属性解缠策略,以实现对具有未知纠缠的新颖属性的编辑。最后,我们扩展了模型以准确编辑细粒度属性。定性和定量评估表明,我们的方法在包括CelebA-HQ、RaFD、Danbooru2018和LSUN教堂在内的各种数据集上优于现有方法。

相似文献

1
FEditNet++: Few-Shot Editing of Latent Semantics in GAN Spaces With Correlated Attribute Disentanglement.FEditNet++:基于相关属性解缠的生成对抗网络空间中潜在语义的少样本编辑
IEEE Trans Pattern Anal Mach Intell. 2024 Dec;46(12):9975-9990. doi: 10.1109/TPAMI.2024.3432529. Epub 2024 Nov 6.
2
Orthogonal Subspace Representation for Generative Adversarial Networks.生成对抗网络的正交子空间表示
IEEE Trans Neural Netw Learn Syst. 2025 Mar;36(3):4413-4427. doi: 10.1109/TNNLS.2024.3377436. Epub 2025 Feb 28.
3
Deep Realistic Facial Editing via Label-restricted Mask Disentanglement.基于标签约束的掩码解缠的深度逼真人脸编辑。
Comput Intell Neurosci. 2022 Nov 23;2022:5652730. doi: 10.1155/2022/5652730. eCollection 2022.
4
Talk-to-Edit: Fine-Grained 2D and 3D Facial Editing via Dialog.通过对话进行精细的二维和三维面部编辑:对话式编辑
IEEE Trans Pattern Anal Mach Intell. 2024 May;46(5):3692-3706. doi: 10.1109/TPAMI.2023.3347299. Epub 2024 Apr 3.
5
Image-to-Image Translation With Disentangled Latent Vectors for Face Editing.用于面部编辑的具有解缠潜向量的图像到图像翻译
IEEE Trans Pattern Anal Mach Intell. 2023 Dec;45(12):14777-14788. doi: 10.1109/TPAMI.2023.3308102. Epub 2023 Nov 3.
6
AttGAN: Facial Attribute Editing by Only Changing What You Want.AttGAN:仅通过改变你想要改变的内容来进行面部属性编辑。
IEEE Trans Image Process. 2019 Nov;28(11):5464-5478. doi: 10.1109/TIP.2019.2916751. Epub 2019 May 20.
7
MaskFaceGAN: High-Resolution Face Editing With Masked GAN Latent Code Optimization.MaskFaceGAN:基于掩码生成对抗网络潜在代码优化的高分辨率面部编辑
IEEE Trans Image Process. 2023;32:5893-5908. doi: 10.1109/TIP.2023.3326675. Epub 2023 Nov 1.
8
Towards Disentangling Latent Space for Unsupervised Semantic Face Editing.面向无监督语义人脸编辑的潜在空间解缠。
IEEE Trans Image Process. 2022;31:1475-1489. doi: 10.1109/TIP.2022.3142527. Epub 2022 Jan 27.
9
GuidedStyle: Attribute knowledge guided style manipulation for semantic face editing.引导式风格:用于语义人脸编辑的属性知识引导式风格操控。
Neural Netw. 2022 Jan;145:209-220. doi: 10.1016/j.neunet.2021.10.017. Epub 2021 Oct 28.
10
InterFaceGAN: Interpreting the Disentangled Face Representation Learned by GANs.InterFaceGAN:解释 GAN 学习到的解缠面部表示。
IEEE Trans Pattern Anal Mach Intell. 2022 Apr;44(4):2004-2018. doi: 10.1109/TPAMI.2020.3034267. Epub 2022 Mar 4.