• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

可扩展的水印技术用于识别大型语言模型输出。

Scalable watermarking for identifying large language model outputs.

机构信息

Google DeepMind, London, UK.

Google, Mountain View, CA, USA.

出版信息

Nature. 2024 Oct;634(8035):818-823. doi: 10.1038/s41586-024-08025-4. Epub 2024 Oct 23.

DOI:10.1038/s41586-024-08025-4
PMID:39443777
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11499265/
Abstract

Large language models (LLMs) have enabled the generation of high-quality synthetic text, often indistinguishable from human-written content, at a scale that can markedly affect the nature of the information ecosystem. Watermarking can help identify synthetic text and limit accidental or deliberate misuse, but has not been adopted in production systems owing to stringent quality, detectability and computational efficiency requirements. Here we describe SynthID-Text, a production-ready text watermarking scheme that preserves text quality and enables high detection accuracy, with minimal latency overhead. SynthID-Text does not affect LLM training and modifies only the sampling procedure; watermark detection is computationally efficient, without using the underlying LLM. To enable watermarking at scale, we develop an algorithm integrating watermarking with speculative sampling, an efficiency technique frequently used in production systems. Evaluations across multiple LLMs empirically show that SynthID-Text provides improved detectability over comparable methods, and standard benchmarks and human side-by-side ratings indicate no change in LLM capabilities. To demonstrate the feasibility of watermarking in large-scale-production systems, we conducted a live experiment that assessed feedback from nearly 20 million Gemini responses, again confirming the preservation of text quality. We hope that the availability of SynthID-Text will facilitate further development of watermarking and responsible use of LLM systems.

摘要

大型语言模型 (LLMs) 能够生成高质量的合成文本,这些文本通常与人类编写的内容难以区分,其规模之大足以显著影响信息生态系统的性质。水印技术可以帮助识别合成文本,并限制意外或故意的滥用,但由于严格的质量、可检测性和计算效率要求,它尚未在生产系统中得到采用。在这里,我们描述了 SynthID-Text,这是一种可用于生产的文本水印方案,它可以在保持文本质量的同时实现高精度的检测,同时具有最小的延迟开销。SynthID-Text 不会影响 LLM 的训练,仅修改采样过程;水印检测具有计算效率,无需使用基础 LLM。为了实现大规模的水印,我们开发了一种将水印与推测性采样相结合的算法,这是一种在生产系统中经常使用的效率技术。在多个 LLM 上的评估经验表明,SynthID-Text 提供了比可比方法更高的可检测性,并且标准基准和人机并排评级表明 LLM 能力没有变化。为了展示水印在大规模生产系统中的可行性,我们进行了一项现场实验,评估了近 2000 万 Gemini 响应的反馈,再次证实了文本质量的保持。我们希望 SynthID-Text 的可用性将促进水印技术的进一步发展和负责任地使用 LLM 系统。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/fd36b16c4ad3/41586_2024_8025_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/406d68054215/41586_2024_8025_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/b9211d58de96/41586_2024_8025_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/95f9782299e2/41586_2024_8025_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/db674a2e581d/41586_2024_8025_Fig4_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/2e9ce789ad24/41586_2024_8025_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/5aae05884342/41586_2024_8025_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/fd36b16c4ad3/41586_2024_8025_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/406d68054215/41586_2024_8025_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/b9211d58de96/41586_2024_8025_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/95f9782299e2/41586_2024_8025_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/db674a2e581d/41586_2024_8025_Fig4_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/2e9ce789ad24/41586_2024_8025_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/5aae05884342/41586_2024_8025_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/ec9d/11499265/fd36b16c4ad3/41586_2024_8025_Fig7_ESM.jpg

相似文献

1
Scalable watermarking for identifying large language model outputs.可扩展的水印技术用于识别大型语言模型输出。
Nature. 2024 Oct;634(8035):818-823. doi: 10.1038/s41586-024-08025-4. Epub 2024 Oct 23.
2
Leveraging Large Language Models for Precision Monitoring of Chemotherapy-Induced Toxicities: A Pilot Study with Expert Comparisons and Future Directions.利用大语言模型进行化疗诱导毒性的精准监测:一项专家比较及未来方向的试点研究
Cancers (Basel). 2024 Aug 12;16(16):2830. doi: 10.3390/cancers16162830.
3
Robust authentication for paper-based text documents based on text watermarking technology.基于文本水印技术的纸质文本文档的稳健认证。
Math Biosci Eng. 2019 Mar 15;16(4):2233-2249. doi: 10.3934/mbe.2019110.
4
Leveraging Large Language Models (LLM) for the Plastic Surgery Resident Training: Do They Have a Role?利用大语言模型进行整形外科住院医师培训:它们能发挥作用吗?
Indian J Plast Surg. 2023 Aug 28;56(5):413-420. doi: 10.1055/s-0043-1772704. eCollection 2023 Oct.
5
Imperceptible reversible watermarking of radiographic images based on quantum noise masking.基于量子噪声掩蔽的射线图像的不可感知可逆水印。
Comput Methods Programs Biomed. 2018 Jul;160:119-128. doi: 10.1016/j.cmpb.2018.03.011. Epub 2018 Mar 21.
6
Large Language Models Can Enable Inductive Thematic Analysis of a Social Media Corpus in a Single Prompt: Human Validation Study.大语言模型可通过单一提示实现社交媒体语料库的归纳主题分析:人类验证研究。
JMIR Infodemiology. 2024 Aug 29;4:e59641. doi: 10.2196/59641.
7
Quality Scalability Aware Watermarking for Visual Content.面向视觉内容的质量可扩展性感知水印技术。
IEEE Trans Image Process. 2016 Nov;25(11):5158-5172. doi: 10.1109/TIP.2016.2599785. Epub 2016 Aug 11.
8
A cloud-based buyer-seller watermarking protocol (CB-BSWP) using semi-trusted third party for copy deterrence and privacy preserving.一种基于云的买卖双方水印协议(CB-BSWP),使用半可信第三方来防止复制并保护隐私。
Multimed Tools Appl. 2022;81(15):21417-21448. doi: 10.1007/s11042-022-12550-7. Epub 2022 Mar 15.
9
Contrasting Linguistic Patterns in Human and LLM-Generated News Text.人类与大语言模型生成的新闻文本中的语言模式对比。
Artif Intell Rev. 2024;57(10):265. doi: 10.1007/s10462-024-10903-2. Epub 2024 Aug 23.
10
Evaluating large language models for health-related text classification tasks with public social media data.利用公共社交媒体数据评估用于健康相关文本分类任务的大型语言模型。
J Am Med Inform Assoc. 2024 Oct 1;31(10):2181-2189. doi: 10.1093/jamia/ocae210.

引用本文的文献

1
Attitudes and perceptions of biomedical journal editors in chief towards the use of artificial intelligence chatbots in the scholarly publishing process: a cross-sectional survey.生物医学期刊主编对学术出版过程中使用人工智能聊天机器人的态度和看法:一项横断面调查
Res Integr Peer Rev. 2025 Sep 8;10(1):19. doi: 10.1186/s41073-025-00178-8.
2
Combating Fake Science in the Age of Generative Artificial Intelligence: A Biomedical Perspective.生成式人工智能时代打击伪科学:生物医学视角
Mo Med. 2025 May-Jun;122(3):162-168.
3
"It is important to consult" a linguist: Verb-Argument Constructions in ChatGPT and human experts' medical and financial advice.

本文引用的文献

1
AI models collapse when trained on recursively generated data.当在递归生成的数据上训练 AI 模型时,模型会崩溃。
Nature. 2024 Jul;631(8022):755-759. doi: 10.1038/s41586-024-07566-y. Epub 2024 Jul 24.
2
GPT detectors are biased against non-native English writers.GPT检测器对非英语母语的写作者存在偏见。
Patterns (N Y). 2023 Jul 10;4(7):100779. doi: 10.1016/j.patter.2023.100779. eCollection 2023 Jul 14.
咨询语言学家很重要:ChatGPT以及人类专家在医学和财务建议方面的动词-论元结构
PLoS One. 2025 May 27;20(5):e0324611. doi: 10.1371/journal.pone.0324611. eCollection 2025.
4
The sociolinguistic foundations of language modeling.语言建模的社会语言学基础。
Front Artif Intell. 2025 Jan 13;7:1472411. doi: 10.3389/frai.2024.1472411. eCollection 2024.
5
Improving authenticity and provenance in digital biomarkers: the case for digital watermarking.提高数字生物标志物的真实性和来源可追溯性:数字水印的应用案例
NPJ Digit Med. 2025 Jan 15;8(1):31. doi: 10.1038/s41746-024-01374-4.
6
FoldMark: Safeguarding Protein Structure Generative Models with Distributional and Evolutionary Watermarking.FoldMark:通过分布和进化水印保护蛋白质结构生成模型
bioRxiv. 2025 Jun 4:2024.10.23.619960. doi: 10.1101/2024.10.23.619960.
7
Google unveils invisible 'watermark' for AI-generated text.谷歌推出用于人工智能生成文本的隐形“水印”。
Nature. 2024 Oct;634(8036):1027-1028. doi: 10.1038/d41586-024-03462-7.