• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

通过使用mT5变压器模型的三阶段微调与强化学习方法来增强波斯语文本摘要。

Enhancing Persian text summarization through a three-phase fine-tuning and reinforcement learning approach with the mT5 transformer model.

作者信息

Abadi Vahid Nejad Mahmood, Ghasemian Fahimeh

机构信息

Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran.

出版信息

Sci Rep. 2025 Jan 2;15(1):80. doi: 10.1038/s41598-024-78235-3.

DOI:10.1038/s41598-024-78235-3
PMID:39747858
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11695816/
Abstract

In the contemporary era, grappling with the vast expanse of big data presents a formidable obstacle, particularly when it comes to extracting vital information from extensive textual sources. The constant influx of news articles from various agencies necessitates an enormous amount of time to digest comprehensively. A viable solution to address this challenge lies in the realm of automatic text summarization, which is a pivotal and intricate endeavor within the field of natural language processing. Text summarization involves transforming pertinent textual content into a concise format that reduces its word count without compromising its underlying meaning. In recent years, transformers have emerged as a prominent force in the landscape of natural language processing, particularly in the realm of text summarization. This research endeavors to harness the power of transformers by training the mT5-base model on a three-step fine-tuning phase on Persian news articles. Subsequently, reinforcement learning via the PPO algorithm is integrated with the fine-tuned model. Finally, we evaluate the model's performance in summarizing Persian texts, shedding light on its efficacy in addressing the formidable task of distilling meaningful insights from a sea of textual data. Our model has set a new benchmark in the field of Persian text summarization, achieving outstanding ROUGE scores of 53.17 for ROUGE-1, 37.12 for ROUGE-2, and 44.13 for ROUGE-L. These remarkable results reflect a significant advancement in the quality of Persian text summarization, signaling a promising era of more refined and context-aware summaries.

摘要

在当代,应对海量大数据是一项艰巨的挑战,尤其是从大量文本来源中提取关键信息时。来自各机构的新闻文章不断涌入,需要大量时间才能全面消化。解决这一挑战的一个可行办法在于自动文本摘要领域,这是自然语言处理领域一项关键且复杂的工作。文本摘要涉及将相关文本内容转换为简洁形式,在不损害其基本含义的情况下减少字数。近年来,Transformer在自然语言处理领域,尤其是在文本摘要领域,已成为一股突出力量。本研究致力于通过在波斯语新闻文章上进行三步微调阶段来训练mT5-base模型,从而利用Transformer的力量。随后,通过近端策略优化(PPO)算法的强化学习与微调后的模型相结合。最后,我们评估该模型在总结波斯语文本方面的性能,揭示其在从海量文本数据中提炼有意义见解这一艰巨任务中的有效性。我们的模型在波斯语文本摘要领域树立了新的标杆,在ROUGE-1上达到了53.17的出色分数,在ROUGE-2上为37.12,在ROUGE-L上为44.13。这些显著成果反映了波斯语文本摘要质量的重大进步,标志着一个更精确、更具上下文感知摘要的充满希望的时代。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25cc/11695816/8cc65ae5741d/41598_2024_78235_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25cc/11695816/1210e8dd1145/41598_2024_78235_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25cc/11695816/8cc65ae5741d/41598_2024_78235_Fig2_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25cc/11695816/1210e8dd1145/41598_2024_78235_Fig1_HTML.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/25cc/11695816/8cc65ae5741d/41598_2024_78235_Fig2_HTML.jpg

相似文献

1
Enhancing Persian text summarization through a three-phase fine-tuning and reinforcement learning approach with the mT5 transformer model.通过使用mT5变压器模型的三阶段微调与强化学习方法来增强波斯语文本摘要。
Sci Rep. 2025 Jan 2;15(1):80. doi: 10.1038/s41598-024-78235-3.
2
Advanced multiple document summarization iterative recursive transformer networks and multimodal transformer.
PeerJ Comput Sci. 2024 Dec 9;10:e2463. doi: 10.7717/peerj-cs.2463. eCollection 2024.
3
CERC: an interactive content extraction, recognition, and construction tool for clinical and biomedical text.CERC:一个用于临床和生物医学文本的交互式内容提取、识别和构建工具。
BMC Med Inform Decis Mak. 2020 Dec 15;20(Suppl 14):306. doi: 10.1186/s12911-020-01330-8.
4
Ontology-based prompt tuning for news article summarization.基于本体的新闻文章摘要提示调整
Front Artif Intell. 2025 Feb 11;8:1520144. doi: 10.3389/frai.2025.1520144. eCollection 2025.
5
Exploring the potential of ChatGPT in medical dialogue summarization: a study on consistency with human preferences.探索 ChatGPT 在医学对话总结中的潜力:一项关于与人类偏好一致性的研究。
BMC Med Inform Decis Mak. 2024 Mar 14;24(1):75. doi: 10.1186/s12911-024-02481-8.
6
Extractive summarization of clinical trial descriptions.临床试验描述的抽取式总结。
Int J Med Inform. 2019 Sep;129:114-121. doi: 10.1016/j.ijmedinf.2019.05.019. Epub 2019 May 30.
7
SATS: simplification aware text summarization of scientific documents.SATS:科学文献的简化感知文本摘要
Front Artif Intell. 2024 Jul 10;7:1375419. doi: 10.3389/frai.2024.1375419. eCollection 2024.
8
Exploring the Efficacy of Large Language Models in Summarizing Mental Health Counseling Sessions: Benchmark Study.探讨大型语言模型在总结心理健康咨询会话中的功效:基准研究。
JMIR Ment Health. 2024 Jul 23;11:e57306. doi: 10.2196/57306.
9
Clinical research text summarization method based on fusion of domain knowledge.基于领域知识融合的临床研究文本摘要方法。
J Biomed Inform. 2024 Aug;156:104668. doi: 10.1016/j.jbi.2024.104668. Epub 2024 Jun 8.
10
Enhancing Radiology Clinical Histories Through Transformer-Based Automated Clinical Note Summarization.通过基于Transformer的自动临床记录摘要增强放射学临床病史
J Imaging Inform Med. 2025 Apr 7. doi: 10.1007/s10278-025-01477-8.

引用本文的文献

1
Hybrid optimization driven fake news detection using reinforced transformer models.基于强化变压器模型的混合优化驱动的假新闻检测
Sci Rep. 2025 Apr 28;15(1):14782. doi: 10.1038/s41598-025-99936-3.