• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

将神经机器翻译扩展到 200 种语言。

Scaling neural machine translation to 200 languages.

出版信息

Nature. 2024 Jun;630(8018):841-846. doi: 10.1038/s41586-024-07335-x. Epub 2024 Jun 5.

DOI:10.1038/s41586-024-07335-x
PMID:38839963
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11208141/
Abstract

The development of neural techniques has opened up new avenues for research in machine translation. Today, neural machine translation (NMT) systems can leverage highly multilingual capacities and even perform zero-shot translation, delivering promising results in terms of language coverage and quality. However, scaling quality NMT requires large volumes of parallel bilingual data, which are not equally available for the 7,000+ languages in the world. Focusing on improving the translation qualities of a relatively small group of high-resource languages comes at the expense of directing research attention to low-resource languages, exacerbating digital inequities in the long run. To break this pattern, here we introduce No Language Left Behind-a single massively multilingual model that leverages transfer learning across languages. We developed a conditional computational model based on the Sparsely Gated Mixture of Experts architecture, which we trained on data obtained with new mining techniques tailored for low-resource languages. Furthermore, we devised multiple architectural and training improvements to counteract overfitting while training on thousands of tasks. We evaluated the performance of our model over 40,000 translation directions using tools created specifically for this purpose-an automatic benchmark (FLORES-200), a human evaluation metric (XSTS) and a toxicity detector that covers every language in our model. Compared with the previous state-of-the-art models, our model achieves an average of 44% improvement in translation quality as measured by BLEU. By demonstrating how to scale NMT to 200 languages and making all contributions in this effort freely available for non-commercial use, our work lays important groundwork for the development of a universal translation system.

摘要

神经技术的发展为机器翻译的研究开辟了新的途径。如今,神经机器翻译 (NMT) 系统可以利用高度多语言能力,甚至可以进行零样本翻译,在语言覆盖范围和质量方面取得了有希望的结果。然而,提高质量 NMT 需要大量的平行双语数据,而这些数据在全球 7000 多种语言中并不是平等可用的。专注于提高相对少数高资源语言的翻译质量,会导致研究注意力转向低资源语言,从长远来看,加剧数字不平等。为了打破这种模式,我们在这里引入了“一个都不能落下”(No Language Left Behind)——一个利用跨语言迁移学习的单一大规模多语言模型。我们基于稀疏门控混合专家(Sparsely Gated Mixture of Experts)架构开发了一个条件计算模型,该模型基于针对低资源语言定制的新挖掘技术进行训练。此外,我们设计了多种架构和训练改进措施,以在数千个任务上进行训练时对抗过拟合。我们使用专门为此目的创建的工具(自动基准测试工具 FLORES-200、人类评估指标 XSTS 和涵盖我们模型中所有语言的毒性检测器),在超过 40000 个翻译方向上评估了我们模型的性能。与之前的最先进模型相比,我们的模型在 BLEU 衡量的翻译质量上平均提高了 44%。通过展示如何将 NMT 扩展到 200 种语言,并免费提供这项工作中的所有贡献供非商业使用,我们的工作为开发通用翻译系统奠定了重要基础。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/072d0ebcdad4/41586_2024_7335_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/df20fc92ef95/41586_2024_7335_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/e7e3b6a95dd2/41586_2024_7335_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/4a305bce5efa/41586_2024_7335_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/5dc0ec104449/41586_2024_7335_Fig4_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/072d0ebcdad4/41586_2024_7335_Fig5_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/df20fc92ef95/41586_2024_7335_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/e7e3b6a95dd2/41586_2024_7335_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/4a305bce5efa/41586_2024_7335_Fig3_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/5dc0ec104449/41586_2024_7335_Fig4_ESM.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/4df2/11208141/072d0ebcdad4/41586_2024_7335_Fig5_ESM.jpg

相似文献

1
Scaling neural machine translation to 200 languages.将神经机器翻译扩展到 200 种语言。
Nature. 2024 Jun;630(8018):841-846. doi: 10.1038/s41586-024-07335-x. Epub 2024 Jun 5.
2
Joint speech and text machine translation for up to 100 languages.支持多达100种语言的联合语音与文本机器翻译。
Nature. 2025 Jan;637(8046):587-593. doi: 10.1038/s41586-024-08359-z. Epub 2025 Jan 15.
3
Neural machine translation of clinical texts between long distance languages.长距离语言之间的临床文本的神经机器翻译。
J Am Med Inform Assoc. 2019 Dec 1;26(12):1478-1487. doi: 10.1093/jamia/ocz110.
4
AmericasNLI: Machine translation and natural language inference systems for Indigenous languages of the Americas.美洲自然语言推理项目:用于美洲原住民语言的机器翻译和自然语言推理系统。
Front Artif Intell. 2022 Dec 2;5:995667. doi: 10.3389/frai.2022.995667. eCollection 2022.
5
Efficient incremental training using a novel NMT-SMT hybrid framework for translation of low-resource languages.使用新颖的神经机器翻译-统计机器翻译混合框架进行低资源语言翻译的高效增量训练。
Front Artif Intell. 2024 Sep 25;7:1381290. doi: 10.3389/frai.2024.1381290. eCollection 2024.
6
Neural machine translation of clinical text: an empirical investigation into multilingual pre-trained language models and transfer-learning.临床文本的神经机器翻译:对多语言预训练语言模型和迁移学习的实证研究。
Front Digit Health. 2024 Feb 26;6:1211564. doi: 10.3389/fdgth.2024.1211564. eCollection 2024.
7
Spanish to Mexican Sign Language glosses corpus for natural language processing tasks.用于自然语言处理任务的西班牙语到墨西哥手语的注释语料库。
Sci Data. 2025 Apr 26;12(1):702. doi: 10.1038/s41597-025-04871-7.
8
Obtaining Parallel Sentences in Low-Resource Language Pairs with Minimal Supervision.用最小监督获取低资源语言对的平行句子。
Comput Intell Neurosci. 2022 Aug 3;2022:5296946. doi: 10.1155/2022/5296946. eCollection 2022.
9
Pseudotext Injection and Advance Filtering of Low-Resource Corpus for Neural Machine Translation.用于神经机器翻译的低资源语料库的伪文本注入与预过滤
Comput Intell Neurosci. 2021 Apr 11;2021:6682385. doi: 10.1155/2021/6682385. eCollection 2021.
10
Adaptation of machine translation for multilingual information retrieval in the medical domain.医学领域中用于多语言信息检索的机器翻译适配
Artif Intell Med. 2014 Jul;61(3):165-85. doi: 10.1016/j.artmed.2014.01.004. Epub 2014 Feb 5.

引用本文的文献

1
The analysis of learning investment effect for artificial intelligence English translation model based on deep neural network.基于深度神经网络的人工智能英语翻译模型学习投资效果分析
Sci Rep. 2025 Jul 19;15(1):26277. doi: 10.1038/s41598-025-11282-6.
2
Large language models in oncology: a review.肿瘤学中的大语言模型:综述
BMJ Oncol. 2025 May 15;4(1):e000759. doi: 10.1136/bmjonc-2025-000759. eCollection 2025.
3
Language-agnostic, Automated Assessment of Listeners' Speech Recall Using Large Language Models.使用大语言模型对听众言语回忆进行与语言无关的自动评估。

本文引用的文献

1
Interdisciplinary Research in Artificial Intelligence: Challenges and Opportunities.人工智能中的跨学科研究:挑战与机遇
Front Big Data. 2020 Nov 23;3:577974. doi: 10.3389/fdata.2020.577974. eCollection 2020.
Trends Hear. 2025 Jan-Dec;29:23312165251347131. doi: 10.1177/23312165251347131. Epub 2025 May 30.
4
A survey of multilingual large language models.多语言大语言模型调查
Patterns (N Y). 2025 Jan 10;6(1):101118. doi: 10.1016/j.patter.2024.101118.
5
Meta AI creates speech-to-speech translator that works in dozens of languages.元人工智能公司开发出了可用于数十种语言的语音到语音翻译器。
Nature. 2025 Jan;637(8047):771-772. doi: 10.1038/d41586-025-00045-y.
6
Demystifying Large Language Models for Medicine: A Primer.揭开医学领域大语言模型的神秘面纱:入门指南。
ArXiv. 2024 Nov 20:arXiv:2410.18856v3.
7
Response accuracy of GPT-4 across languages: insights from an expert-level diagnostic radiology examination in Japan.GPT-4在多种语言中的回答准确性:来自日本专家级诊断放射学考试的见解。
Jpn J Radiol. 2025 Feb;43(2):319-329. doi: 10.1007/s11604-024-01673-6. Epub 2024 Oct 28.