• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于动画的在线课堂环境中的面部图像操作的实验研究。

An experimental study of animating-based facial image manipulation in online class environments.

机构信息

Graduate School of Data Science, Seoul National University of Science and Technology, Seoul, South Korea.

Department of Industrial Engineering, Seoul National University of Science and Technology, Seoul, South Korea.

出版信息

Sci Rep. 2023 Mar 22;13(1):4667. doi: 10.1038/s41598-023-31408-y.

DOI:10.1038/s41598-023-31408-y
PMID:36949178
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10031167/
Abstract

Recent advances in artificial intelligence technology have significantly improved facial image manipulation, which is known as Deepfake. Facial image manipulation synthesizes or replaces a region of the face in an image with that of another face. The techniques for facial image manipulation are classified into four categories: (1) entire face synthesis, (2) identity swap, (3) attribute manipulation, and (4) expression swap. Out of them, we focus on expression swap because it effectively manipulates only the expression of the face in the images or videos without creating or replacing the entire face, having advantages for the real-time application. In this study, we propose an evaluation framework of the expression swap models targeting the real-time online class environments. For this, we define three kinds of scenarios according to the portion of the face in the entire image considering actual online class situations: (1) attendance check (Scenario 1), (2) presentation (Scenario 2), and (3) examination (Scenario 3). Considering the manipulation on the online class environments, the framework receives a single source image and a target video and generates the video that manipulates a face of the target video to that in the source image. To this end, we select two models that satisfy the conditions required by the framework: (1) first order model and (2) GANimation. We implement these models in the framework and evaluate their performance for the defined scenarios. Through the quantitative and qualitative evaluation, we observe distinguishing properties of the used two models. Specifically, both models show acceptable results in Scenario 1, where the face occupies a large portion of the image. However, their performances are significantly degraded in Scenarios 2 and 3, where the face occupies less portion of the image; the first order model causes relatively less loss of image quality than GANimation in the result of the quantitative evaluation. In contrast, GANimation has the advantages of representing facial expression changes compared to the first order model. Finally, we devise an architecture for applying the expression swap model to the online video conferencing application in real-time. In particular, by applying the expression swap model to widely used online meeting platforms such as Zoom, Google Meet, and Microsoft Teams, we demonstrate its feasibility for real-time online classes.

摘要

最近,人工智能技术的进步显著提高了人脸图像操纵技术,即 Deepfake。人脸图像操纵通过将图像中某一区域的人脸替换为另一张人脸来实现。人脸图像操纵技术主要分为四类:(1)整体面部合成,(2)身份交换,(3)属性操作,(4)表情交换。其中,我们重点关注表情交换,因为它可以有效地仅操纵图像或视频中的面部表情,而无需创建或替换整个面部,这对于实时应用具有优势。在这项研究中,我们针对实时在线课堂环境提出了一种表情交换模型的评估框架。为此,我们根据考虑实际在线课堂情况的整个图像中的面部部分定义了三种情况:(1)出勤检查(场景 1),(2)演示(场景 2)和(3)考试(场景 3)。考虑到对在线课堂环境的操纵,该框架接收单张源图像和目标视频,并生成操纵目标视频中的面部以使其与源图像中的面部相同的视频。为此,我们选择了两个满足框架要求的模型:(1)一阶模型和(2)GANimation。我们在框架中实现了这些模型,并针对定义的场景评估了它们的性能。通过定量和定性评估,我们观察到所使用的两个模型的区别特征。具体来说,两个模型在场景 1 中表现良好,其中面部占据图像的大部分。但是,在场景 2 和 3 中,当面部占据图像的较小部分时,它们的性能会显著下降;在定量评估的结果中,一阶模型导致的图像质量损失相对较小。相比之下,GANimation 具有比一阶模型更好地表示面部表情变化的优势。最后,我们设计了一种架构,可将表情交换模型实时应用于在线视频会议应用程序中。特别是,通过将表情交换模型应用于广泛使用的在线会议平台,如 Zoom、Google Meet 和 Microsoft Teams,我们展示了其在实时在线课程中的可行性。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/aca0e4239e0c/41598_2023_31408_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/07aeb01d2784/41598_2023_31408_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/3faa6afd63d9/41598_2023_31408_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/179d8d4f3450/41598_2023_31408_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/6a170742d0ce/41598_2023_31408_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/f29e8b12f7fe/41598_2023_31408_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/42bc1c2367bb/41598_2023_31408_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/36ce64b1fdcb/41598_2023_31408_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/f34238339aa6/41598_2023_31408_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/dbd6c61e7ed4/41598_2023_31408_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/9f0066e4d685/41598_2023_31408_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/aca0e4239e0c/41598_2023_31408_Fig11_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/07aeb01d2784/41598_2023_31408_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/3faa6afd63d9/41598_2023_31408_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/179d8d4f3450/41598_2023_31408_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/6a170742d0ce/41598_2023_31408_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/f29e8b12f7fe/41598_2023_31408_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/42bc1c2367bb/41598_2023_31408_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/36ce64b1fdcb/41598_2023_31408_Fig7_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/f34238339aa6/41598_2023_31408_Fig8_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/dbd6c61e7ed4/41598_2023_31408_Fig9_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/9f0066e4d685/41598_2023_31408_Fig10_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/dec0/10033672/aca0e4239e0c/41598_2023_31408_Fig11_HTML.jpg

相似文献

1
An experimental study of animating-based facial image manipulation in online class environments.基于动画的在线课堂环境中的面部图像操作的实验研究。
Sci Rep. 2023 Mar 22;13(1):4667. doi: 10.1038/s41598-023-31408-y.
2
Do Deepfakes Adequately Display Emotions? A Study on Deepfake Facial Emotion Expression.深度伪造能否充分展示情感?一项关于深度伪造面部表情表达的研究。
Comput Intell Neurosci. 2022 Oct 18;2022:1332122. doi: 10.1155/2022/1332122. eCollection 2022.
3
A Review of Image Processing Techniques for Deepfakes.深度伪造的图像处理技术综述。
Sensors (Basel). 2022 Jun 16;22(12):4556. doi: 10.3390/s22124556.
4
[Standard technical specifications for methacholine chloride (Methacholine) bronchial challenge test (2023)].[氯化乙酰甲胆碱支气管激发试验标准技术规范(2023年)]
Zhonghua Jie He He Hu Xi Za Zhi. 2024 Feb 12;47(2):101-119. doi: 10.3760/cma.j.cn112147-20231019-00247.
5
How Can Research on Artificial Empathy Be Enhanced by Applying Deepfakes?如何通过应用深度伪造技术来加强对人工同理心的研究?
J Med Internet Res. 2022 Mar 4;24(3):e29506. doi: 10.2196/29506.
6
Deepfakes Generation and Detection: A Short Survey.深度伪造的生成与检测:简要综述
J Imaging. 2023 Jan 13;9(1):18. doi: 10.3390/jimaging9010018.
7
A Review on Automatic Facial Expression Recognition Systems Assisted by Multimodal Sensor Data.基于多模态传感器数据的自动面部表情识别系统综述。
Sensors (Basel). 2019 Apr 18;19(8):1863. doi: 10.3390/s19081863.
8
Dynamics of facial expression: recognition of facial actions and their temporal segments from face profile image sequences.面部表情动态:从面部轮廓图像序列中识别面部动作及其时间片段。
IEEE Trans Syst Man Cybern B Cybern. 2006 Apr;36(2):433-49. doi: 10.1109/tsmcb.2005.859075.
9
Hierarchical Spatio-Temporal Probabilistic Graphical Model with Multiple Feature Fusion for Binary Facial Attribute Classification in Real-World Face Videos.用于真实人脸视频中二元人脸属性分类的具有多特征融合的分层时空概率图模型。
IEEE Trans Pattern Anal Mach Intell. 2016 Jun;38(6):1185-203. doi: 10.1109/TPAMI.2015.2481396. Epub 2015 Sep 23.
10
Active and dynamic information fusion for facial expression understanding from image sequences.用于从图像序列理解面部表情的主动动态信息融合
IEEE Trans Pattern Anal Mach Intell. 2005 May;27(5):699-714. doi: 10.1109/TPAMI.2005.93.

引用本文的文献

1
GBMPhos: A Gating Mechanism and Bi-GRU-Based Method for Identifying Phosphorylation Sites of SARS-CoV-2 Infection.GBMPhos:一种基于门控机制和双向门控循环单元的新冠病毒感染磷酸化位点识别方法。
Biology (Basel). 2024 Oct 6;13(10):798. doi: 10.3390/biology13100798.
2
Deepfake forensics: a survey of digital forensic methods for multimodal deepfake identification on social media.深度伪造取证:社交媒体上多模态深度伪造识别的数字取证方法综述
PeerJ Comput Sci. 2024 May 27;10:e2037. doi: 10.7717/peerj-cs.2037. eCollection 2024.

本文引用的文献

1
GuidedStyle: Attribute knowledge guided style manipulation for semantic face editing.引导式风格:用于语义人脸编辑的属性知识引导式风格操控。
Neural Netw. 2022 Jan;145:209-220. doi: 10.1016/j.neunet.2021.10.017. Epub 2021 Oct 28.
2
A Style-Based Generator Architecture for Generative Adversarial Networks.基于风格的生成对抗网络生成器架构。
IEEE Trans Pattern Anal Mach Intell. 2021 Dec;43(12):4217-4228. doi: 10.1109/TPAMI.2020.2970919. Epub 2021 Nov 3.
3
GANimation: Anatomically-aware Facial Animation from a Single Image.GANimation:基于单张图像的解剖学感知面部动画
Comput Vis ECCV. 2018 Sep;11214:835-851. doi: 10.1007/978-3-030-01249-6_50. Epub 2018 Oct 6.
4
Image quality assessment: from error visibility to structural similarity.图像质量评估:从误差可见性到结构相似性。
IEEE Trans Image Process. 2004 Apr;13(4):600-12. doi: 10.1109/tip.2003.819861.