• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

更大且更具指导性的语言模型变得不那么可靠。

Larger and more instructable language models become less reliable.

作者信息

Zhou Lexin, Schellaert Wout, Martínez-Plumed Fernando, Moros-Daval Yael, Ferri Cèsar, Hernández-Orallo José

机构信息

Valencian Research Institute for Artificial Intelligence (VRAIN), Universitat Politècnica de València, Valencia, Spain.

University of Cambridge, Cambridge, UK.

出版信息

Nature. 2024 Oct;634(8032):61-68. doi: 10.1038/s41586-024-07930-y. Epub 2024 Sep 25.

DOI:10.1038/s41586-024-07930-y
PMID:39322679
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11446866/
Abstract

The prevailing methods to make large language models more powerful and amenable have been based on continuous scaling up (that is, increasing their size, data volume and computational resources) and bespoke shaping up (including post-filtering, fine tuning or use of human feedback). However, larger and more instructable large language models may have become less reliable. By studying the relationship between difficulty concordance, task avoidance and prompting stability of several language model families, here we show that easy instances for human participants are also easy for the models, but scaled-up, shaped-up models do not secure areas of low difficulty in which either the model does not err or human supervision can spot the errors. We also find that early models often avoid user questions but scaled-up, shaped-up models tend to give an apparently sensible yet wrong answer much more often, including errors on difficult questions that human supervisors frequently overlook. Moreover, we observe that stability to different natural phrasings of the same question is improved by scaling-up and shaping-up interventions, but pockets of variability persist across difficulty levels. These findings highlight the need for a fundamental shift in the design and development of general-purpose artificial intelligence, particularly in high-stakes areas for which a predictable distribution of errors is paramount.

摘要

使大语言模型更强大且更易于使用的主流方法一直基于持续扩大规模(即增加其大小、数据量和计算资源)以及定制塑造(包括后过滤、微调或使用人工反馈)。然而,更大且更具指令性的大语言模型可能变得不那么可靠了。通过研究几个语言模型家族的难度一致性、任务回避和提示稳定性之间的关系,我们在此表明,对人类参与者来说简单的实例对模型来说也简单,但扩大规模、塑造后的模型并不能确保存在低难度区域,在这些区域中模型不会出错或人工监督能够发现错误。我们还发现,早期模型常常回避用户问题,但扩大规模、塑造后的模型更倾向于给出看似合理却错误的答案,包括在人类监督者经常忽略的难题上出现的错误。此外,我们观察到,通过扩大规模和塑造干预,对同一问题不同自然表述的稳定性得到了提高,但在不同难度级别上仍存在一些变异性。这些发现凸显了在通用人工智能的设计和开发中进行根本性转变的必要性,尤其是在对错误的可预测分布至关重要的高风险领域。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/0f197e6999ed/41586_2024_7930_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/1bd506e0c15c/41586_2024_7930_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/d1a039d90d75/41586_2024_7930_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/c6238269b224/41586_2024_7930_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/35ebb73d66b2/41586_2024_7930_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/17f049dbde74/41586_2024_7930_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/fca993edf241/41586_2024_7930_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/b2df1d3dfee8/41586_2024_7930_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/0f197e6999ed/41586_2024_7930_Fig8_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/1bd506e0c15c/41586_2024_7930_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/d1a039d90d75/41586_2024_7930_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/c6238269b224/41586_2024_7930_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/35ebb73d66b2/41586_2024_7930_Fig4_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/17f049dbde74/41586_2024_7930_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/fca993edf241/41586_2024_7930_Fig6_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/b2df1d3dfee8/41586_2024_7930_Fig7_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/a8b3/11446866/0f197e6999ed/41586_2024_7930_Fig8_ESM.jpg

相似文献

1
Larger and more instructable language models become less reliable.更大且更具指导性的语言模型变得不那么可靠。
Nature. 2024 Oct;634(8032):61-68. doi: 10.1038/s41586-024-07930-y. Epub 2024 Sep 25.
2
Folic acid supplementation and malaria susceptibility and severity among people taking antifolate antimalarial drugs in endemic areas.在流行地区,服用抗叶酸抗疟药物的人群中,叶酸补充剂与疟疾易感性和严重程度的关系。
Cochrane Database Syst Rev. 2022 Feb 1;2(2022):CD014217. doi: 10.1002/14651858.CD014217.
3
Macromolecular crowding: chemistry and physics meet biology (Ascona, Switzerland, 10-14 June 2012).大分子拥挤现象:化学与物理邂逅生物学(瑞士阿斯科纳,2012年6月10日至14日)
Phys Biol. 2013 Aug;10(4):040301. doi: 10.1088/1478-3975/10/4/040301. Epub 2013 Aug 2.
4
Fine-tuning protein language models boosts predictions across diverse tasks.微调蛋白质语言模型可提高跨多种任务的预测能力。
Nat Commun. 2024 Aug 28;15(1):7407. doi: 10.1038/s41467-024-51844-2.
5
Reshaping free-text radiology notes into structured reports with generative question answering transformers.利用生成式问答变换模型将自由文本放射学报告改造成结构化报告。
Artif Intell Med. 2024 Aug;154:102924. doi: 10.1016/j.artmed.2024.102924. Epub 2024 Jun 26.
6
A Language Model-Powered Simulated Patient With Automated Feedback for History Taking: Prospective Study.基于语言模型的模拟患者与自动化反馈的病史采集:前瞻性研究。
JMIR Med Educ. 2024 Aug 16;10:e59213. doi: 10.2196/59213.
7
Model tuning or prompt Tuning? a study of large language models for clinical concept and relation extraction.模型调优还是提示调优?大型语言模型在临床概念和关系抽取中的应用研究。
J Biomed Inform. 2024 May;153:104630. doi: 10.1016/j.jbi.2024.104630. Epub 2024 Mar 26.
8
A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.从自由文本肿瘤病理学报告(CancerBERT 网络)中提取数据的问答系统:开发研究。
J Med Internet Res. 2022 Mar 23;24(3):e27210. doi: 10.2196/27210.
9
A large language model-based generative natural language processing framework fine-tuned on clinical notes accurately extracts headache frequency from electronic health records.基于大型语言模型的生成式自然语言处理框架,在临床笔记上进行了微调,能够从电子健康记录中准确提取头痛频率。
Headache. 2024 Apr;64(4):400-409. doi: 10.1111/head.14702. Epub 2024 Mar 25.
10
Unlocking the Secrets Behind Advanced Artificial Intelligence Language Models in Deidentifying Chinese-English Mixed Clinical Text: Development and Validation Study.揭开高级人工智能语言模型在去识别汉英混合临床文本背后的秘密:开发与验证研究。
J Med Internet Res. 2024 Jan 25;26:e48443. doi: 10.2196/48443.

引用本文的文献

1
Synthetic Scientific Image Generation with VAE, GAN, and Diffusion Model Architectures.使用变分自编码器(VAE)、生成对抗网络(GAN)和扩散模型架构进行合成科学图像生成。
J Imaging. 2025 Jul 26;11(8):252. doi: 10.3390/jimaging11080252.
2
Artificial Intelligence in migrant health: a critical perspective on opportunities and risks.人工智能在移民健康中的应用:对机遇与风险的批判性视角
Lancet Reg Health Eur. 2025 Aug 8;57:101421. doi: 10.1016/j.lanepe.2025.101421. eCollection 2025 Oct.
3
Optimizing patient education for radioactive iodine therapy and the role of ChatGPT incorporating chain-of-thought technique: ChatGPT questionnaire.

本文引用的文献

1
Artificial intelligence and illusions of understanding in scientific research.人工智能与科研中的理解错觉。
Nature. 2024 Mar;627(8002):49-58. doi: 10.1038/s41586-024-07146-0. Epub 2024 Mar 6.
2
Autonomous chemical research with large language models.大语言模型驱动的自主化学研究。
Nature. 2023 Dec;624(7992):570-578. doi: 10.1038/s41586-023-06792-0. Epub 2023 Dec 20.
3
Large language models in medicine.医学中的大型语言模型。
优化放射性碘治疗的患者教育以及结合思维链技术的ChatGPT的作用:ChatGPT问卷
Digit Health. 2025 Jul 7;11:20552076251357468. doi: 10.1177/20552076251357468. eCollection 2025 Jan-Dec.
4
Take caution in using LLMs as human surrogates.谨慎使用大型语言模型作为人类替代者。
Proc Natl Acad Sci U S A. 2025 Jun 17;122(24):e2501660122. doi: 10.1073/pnas.2501660122. Epub 2025 Jun 13.
5
Patient agency and large language models in worldwide encoding of equity.全球公平编码中的患者能动性与大语言模型
NPJ Digit Med. 2025 May 8;8(1):258. doi: 10.1038/s41746-025-01598-y.
6
Mitigating the risk of health inequity exacerbated by large language models.减轻大语言模型加剧的健康不平等风险。
NPJ Digit Med. 2025 May 4;8(1):246. doi: 10.1038/s41746-025-01576-4.
7
Generalization bias in large language model summarization of scientific research.大语言模型对科学研究进行总结时的泛化偏差。
R Soc Open Sci. 2025 Apr 30;12(4):241776. doi: 10.1098/rsos.241776. eCollection 2025 Apr.
8
ShennongAlpha: an AI-driven sharing and collaboration platform for intelligent curation, acquisition, and translation of natural medicinal material knowledge.神农α:一个由人工智能驱动的共享与协作平台,用于天然药材知识的智能整理、获取及翻译。
Cell Discov. 2025 Apr 1;11(1):32. doi: 10.1038/s41421-025-00776-2.
9
Cross sectional pilot study on clinical review generation using large language models.使用大语言模型生成临床综述的横断面试点研究。
NPJ Digit Med. 2025 Mar 19;8(1):170. doi: 10.1038/s41746-025-01535-z.
10
Agents for Change: Artificial Intelligent Workflows for Quantitative Clinical Pharmacology and Translational Sciences.变革推动者:定量临床药理学与转化科学的人工智能工作流程
Clin Transl Sci. 2025 Mar;18(3):e70188. doi: 10.1111/cts.70188.
Nat Med. 2023 Aug;29(8):1930-1940. doi: 10.1038/s41591-023-02448-8. Epub 2023 Jul 17.
4
General intelligence disentangled via a generality metric for natural and artificial intelligence.通过自然和人工智能的一般性度量来分解通用智能。
Sci Rep. 2021 Nov 24;11(1):22822. doi: 10.1038/s41598-021-01997-7.
5
Case-mixing effects on anagram solution.病例组合对变位词解谜的影响。
J Gen Psychol. 2002 Apr;129(2):117-26. doi: 10.1080/00221300209603132.
6
Solution of anagrams.变位词的解法。
Psychol Bull. 1966 Nov;66(5):371-84. doi: 10.1037/h0023886.