• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

UniFaceGAN:一种用于时间一致的面部视频编辑的统一框架。

UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing.

出版信息

IEEE Trans Image Process. 2021;30:6107-6116. doi: 10.1109/TIP.2021.3089909. Epub 2021 Jul 7.

DOI:10.1109/TIP.2021.3089909
PMID:34166189
Abstract

Recent research has witnessed advances in facial image editing tasks including face swapping and face reenactment. However, these methods are confined to dealing with one specific task at a time. In addition, for video facial editing, previous methods either simply apply transformations frame by frame or utilize multiple frames in a concatenated or iterative fashion, which leads to noticeable visual flickers. In this paper, we propose a unified temporally consistent facial video editing framework termed UniFaceGAN. Based on a 3D reconstruction model and a simple yet efficient dynamic training sample selection mechanism, our framework is designed to handle face swapping and face reenactment simultaneously. To enforce the temporal consistency, a novel 3D temporal loss constraint is introduced based on the barycentric coordinate interpolation. Besides, we propose a region-aware conditional normalization layer to replace the traditional AdaIN or SPADE to synthesize more context-harmonious results. Compared with the state-of-the-art facial image editing methods, our framework generates video portraits that are more photo-realistic and temporally smooth.

摘要

最近的研究在面部图像编辑任务方面取得了进展,包括换脸和人脸重放。然而,这些方法一次只能处理一个特定的任务。此外,对于视频面部编辑,以前的方法要么简单地逐帧应用变换,要么以串联或迭代的方式使用多个帧,这导致明显的视觉闪烁。在本文中,我们提出了一种统一的、时间一致的面部视频编辑框架,称为 UniFaceGAN。基于 3D 重建模型和一个简单而高效的动态训练样本选择机制,我们的框架旨在同时处理换脸和人脸重放。为了强制实现时间一致性,我们根据重心坐标插值引入了一种新颖的 3D 时间损失约束。此外,我们提出了一种基于区域感知的条件归一化层来替代传统的 AdaIN 或 SPADE,以生成更具有上下文协调性的结果。与最先进的面部图像编辑方法相比,我们的框架生成的视频人像更加逼真和时间平滑。

相似文献

1
UniFaceGAN: A Unified Framework for Temporally Consistent Facial Video Editing.UniFaceGAN:一种用于时间一致的面部视频编辑的统一框架。
IEEE Trans Image Process. 2021;30:6107-6116. doi: 10.1109/TIP.2021.3089909. Epub 2021 Jul 7.
2
FSGANv2: Improved Subject Agnostic Face Swapping and Reenactment.FSGANv2:改进的与主体无关的面部交换和重演
IEEE Trans Pattern Anal Mach Intell. 2023 Jan;45(1):560-575. doi: 10.1109/TPAMI.2022.3155571. Epub 2022 Dec 5.
3
AttGAN: Facial Attribute Editing by Only Changing What You Want.AttGAN:仅通过改变你想要改变的内容来进行面部属性编辑。
IEEE Trans Image Process. 2019 Nov;28(11):5464-5478. doi: 10.1109/TIP.2019.2916751. Epub 2019 May 20.
4
Video Frame Interpolation and Enhancement via Pyramid Recurrent Framework.基于金字塔循环框架的视频帧插值与增强
IEEE Trans Image Process. 2021;30:277-292. doi: 10.1109/TIP.2020.3033617. Epub 2020 Nov 20.
5
Identity-Aware and Shape-Aware Propagation of Face Editing in Videos.视频中面部编辑的身份感知与形状感知传播
IEEE Trans Vis Comput Graph. 2024 Jul;30(7):3444-3456. doi: 10.1109/TVCG.2023.3235364. Epub 2024 Jun 27.
6
A Temporally-Aware Interpolation Network for Video Frame Inpainting.用于视频帧修复的时间感知插值网络
IEEE Trans Pattern Anal Mach Intell. 2020 May;42(5):1053-1068. doi: 10.1109/TPAMI.2019.2951667. Epub 2019 Nov 6.
7
Deep Realistic Facial Editing via Label-restricted Mask Disentanglement.基于标签约束的掩码解缠的深度逼真人脸编辑。
Comput Intell Neurosci. 2022 Nov 23;2022:5652730. doi: 10.1155/2022/5652730. eCollection 2022.
8
Temporally Coherent Video Harmonization Using Adversarial Networks.基于对抗网络的时域一致视频协调
IEEE Trans Image Process. 2020;29:214-224. doi: 10.1109/TIP.2019.2925550. Epub 2019 Jul 17.
9
F³A-GAN: Facial Flow for Face Animation With Generative Adversarial Networks.F³A-GAN:基于生成对抗网络的人脸动画的面部流
IEEE Trans Image Process. 2021;30:8658-8670. doi: 10.1109/TIP.2021.3112059. Epub 2021 Oct 21.
10
FACEMUG: A Multimodal Generative and Fusion Framework for Local Facial Editing.FACEMUG:一种用于局部面部编辑的多模态生成与融合框架。
IEEE Trans Vis Comput Graph. 2024 Jul 26;PP. doi: 10.1109/TVCG.2024.3434386.

引用本文的文献

1
Deepfakes Generation and Detection: A Short Survey.深度伪造的生成与检测:简要综述
J Imaging. 2023 Jan 13;9(1):18. doi: 10.3390/jimaging9010018.