• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

跨转录本、音频和视频检测政治演讲深度伪造。

Human detection of political speech deepfakes across transcripts, audio, and video.

机构信息

Kellogg School of Management, Northwestern University, Evanston, IL, USA.

Media Lab, Massachusetts Institute of Technology, Cambridge, MA, USA.

出版信息

Nat Commun. 2024 Sep 2;15(1):7629. doi: 10.1038/s41467-024-51998-z.

DOI:10.1038/s41467-024-51998-z
PMID:39223110
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11368926/
Abstract

Recent advances in technology for hyper-realistic visual and audio effects provoke the concern that deepfake videos of political speeches will soon be indistinguishable from authentic video. We conduct 5 pre-registered randomized experiments with N = 2215 participants to evaluate how accurately humans distinguish real political speeches from fabrications across base rates of misinformation, audio sources, question framings with and without priming, and media modalities. We do not find base rates of misinformation have statistically significant effects on discernment. We find deepfakes with audio produced by the state-of-the-art text-to-speech algorithms are harder to discern than the same deepfakes with voice actor audio. Moreover across all experiments and question framings, we find audio and visual information enables more accurate discernment than text alone: human discernment relies more on how something is said, the audio-visual cues, than what is said, the speech content.

摘要

最近在超逼真视觉和音频效果方面的技术进步引发了人们的担忧,即政治演讲的深度伪造视频将很快与真实视频无法区分。我们进行了 5 项预先注册的随机实验,共有 2215 名参与者,以评估人类在各种错误信息基数、音频来源、带有和不带有启动的问题框架以及媒体模式下,对真实政治演讲和伪造演讲的准确区分程度。我们没有发现错误信息的基本比率对识别有统计学上的显著影响。我们发现,由最先进的文本到语音算法生成的音频的深度伪造比具有语音演员音频的深度伪造更难识别。此外,在所有实验和问题框架中,我们发现音频和视觉信息比仅文本更能准确识别:人类识别更多地依赖于说话的方式,即音频-视觉线索,而不是所说的内容,即演讲内容。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/5b313871e912/41467_2024_51998_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/ed446bd76ccf/41467_2024_51998_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/87ec063fce2f/41467_2024_51998_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/cb0bb2fde44f/41467_2024_51998_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/8036fa4fc1ab/41467_2024_51998_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/db0ae323e57d/41467_2024_51998_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/5b313871e912/41467_2024_51998_Fig6_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/ed446bd76ccf/41467_2024_51998_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/87ec063fce2f/41467_2024_51998_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/cb0bb2fde44f/41467_2024_51998_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/8036fa4fc1ab/41467_2024_51998_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/db0ae323e57d/41467_2024_51998_Fig5_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d8a9/11368926/5b313871e912/41467_2024_51998_Fig6_HTML.jpg

相似文献

1
Human detection of political speech deepfakes across transcripts, audio, and video.跨转录本、音频和视频检测政治演讲深度伪造。
Nat Commun. 2024 Sep 2;15(1):7629. doi: 10.1038/s41467-024-51998-z.
2
Audio deepfakes: A survey.音频深度伪造:一项调查。
Front Big Data. 2023 Jan 9;5:1001063. doi: 10.3389/fdata.2022.1001063. eCollection 2022.
3
Investigation of Deepfake Voice Detection Using Speech Pause Patterns: Algorithm Development and Validation.基于语音停顿模式的深度伪造语音检测研究:算法开发与验证
JMIR Biomed Eng. 2024 Mar 21;9:e56245. doi: 10.2196/56245.
4
A Robust Approach to Multimodal Deepfake Detection.一种用于多模态深度伪造检测的稳健方法。
J Imaging. 2023 Jun 19;9(6):122. doi: 10.3390/jimaging9060122.
5
A Review of Image Processing Techniques for Deepfakes.深度伪造的图像处理技术综述。
Sensors (Basel). 2022 Jun 16;22(12):4556. doi: 10.3390/s22124556.
6
Do Deepfakes Adequately Display Emotions? A Study on Deepfake Facial Emotion Expression.深度伪造能否充分展示情感?一项关于深度伪造面部表情表达的研究。
Comput Intell Neurosci. 2022 Oct 18;2022:1332122. doi: 10.1155/2022/1332122. eCollection 2022.
7
Deepfake detection by human crowds, machines, and machine-informed crowds.基于人类群体、机器和机器辅助的人类群体的深度伪造检测。
Proc Natl Acad Sci U S A. 2022 Jan 4;119(1). doi: 10.1073/pnas.2110013119.
8
Warning: Humans cannot reliably detect speech deepfakes.警告:人类无法可靠地识别语音深度伪造。
PLoS One. 2023 Aug 2;18(8):e0285333. doi: 10.1371/journal.pone.0285333. eCollection 2023.
9
Do deepfake videos undermine our epistemic trust? A thematic analysis of tweets that discuss deepfakes in the Russian invasion of Ukraine.深度伪造视频是否破坏了我们的认知信任?对讨论俄罗斯入侵乌克兰中的深度伪造的推文的主题分析。
PLoS One. 2023 Oct 25;18(10):e0291668. doi: 10.1371/journal.pone.0291668. eCollection 2023.
10
Deepfakes as a threat to a speaker and facial recognition: An overview of tools and attack vectors.深度伪造对说话者和面部识别的威胁:工具与攻击途径概述
Heliyon. 2023 Apr 3;9(4):e15090. doi: 10.1016/j.heliyon.2023.e15090. eCollection 2023 Apr.

引用本文的文献

1
Generative propaganda: Evidence of AI's impact from a state-backed disinformation campaign.生成式宣传:来自一场国家支持的虚假信息运动的人工智能影响证据。
PNAS Nexus. 2025 Apr 1;4(4):pgaf083. doi: 10.1093/pnasnexus/pgaf083. eCollection 2025 Apr.
2
People are poorly equipped to detect AI-powered voice clones.人们识别人工智能驱动的语音克隆的能力很差。
Sci Rep. 2025 Mar 31;15(1):11004. doi: 10.1038/s41598-025-94170-3.

本文引用的文献

1
Misunderstanding the harms of online misinformation.误解网络错误信息的危害。
Nature. 2024 Jun;630(8015):45-53. doi: 10.1038/s41586-024-07417-w. Epub 2024 Jun 5.
2
Understanding and combatting misinformation across 16 countries on six continents.理解并打击六大洲 16 个国家的错误信息。
Nat Hum Behav. 2023 Sep;7(9):1502-1513. doi: 10.1038/s41562-023-01641-6. Epub 2023 Jun 29.
3
Art and the science of generative AI.生成式人工智能的艺术与科学。
Science. 2023 Jun 16;380(6650):1110-1111. doi: 10.1126/science.adh4451. Epub 2023 Jun 15.
4
The social media context interferes with truth discernment.社交媒体语境会干扰真相识别。
Sci Adv. 2023 Mar 3;9(9):eabo6169. doi: 10.1126/sciadv.abo6169.
5
Protecting world leaders against deep fakes using facial, gestural, and vocal mannerisms.利用面部、手势和声音习惯特征保护世界领导人免受深度伪造技术的影响。
Proc Natl Acad Sci U S A. 2022 Nov 29;119(48):e2216035119. doi: 10.1073/pnas.2216035119. Epub 2022 Nov 23.
6
Psychological inoculation improves resilience against misinformation on social media.心理预接种可提高抵御社交媒体错误信息的恢复力。
Sci Adv. 2022 Aug 26;8(34):eabo6254. doi: 10.1126/sciadv.abo6254. Epub 2022 Aug 24.
7
AI-synthesized faces are indistinguishable from real faces and more trustworthy.人工智能合成的人脸与真实人脸难以区分,并且更值得信任。
Proc Natl Acad Sci U S A. 2022 Feb 22;119(8). doi: 10.1073/pnas.2120481119.
8
Deepfake detection by human crowds, machines, and machine-informed crowds.基于人类群体、机器和机器辅助的人类群体的深度伪造检测。
Proc Natl Acad Sci U S A. 2022 Jan 4;119(1). doi: 10.1073/pnas.2110013119.
9
How do you solve a problem like misinformation?你如何解决像错误信息这样的问题?
Sci Adv. 2021 Dec 10;7(50):eabn0481. doi: 10.1126/sciadv.abn0481. Epub 2021 Dec 8.
10
Fooled twice: People cannot detect deepfakes but think they can.被愚弄两次:人们无法察觉深度伪造,但却认为自己能做到。
iScience. 2021 Oct 29;24(11):103364. doi: 10.1016/j.isci.2021.103364. eCollection 2021 Nov 19.