• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于摘要输出区域的用于长度控制的抽象摘要的增强型变压器。

Enhanced transformer for length-controlled abstractive summarization based on summary output area.

作者信息

Sunusi Yusuf, Omar Nazlia, Zakaria Lailatul Qadri

机构信息

Center for Artificial Intelligence Technology, Faculty of Information Science and Technology, Universiti Kebangsaan Malaysia, Bangi, Selangor, Malaysia.

出版信息

PeerJ Comput Sci. 2025 Mar 11;11:e2667. doi: 10.7717/peerj-cs.2667. eCollection 2025.

DOI:10.7717/peerj-cs.2667
PMID:40134863
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11935775/
Abstract

Recent advancements in abstractive summarization models, particularly those built on encoder-decoder architectures, typically produce a single summary for each source text. Controlling the length of summaries is crucial for practical applications, such as crafting cover summaries for newspapers or magazines with varying slot sizes. Current research in length-controllable abstractive summarization employs techniques like length embeddings in the decoder module or a word-level extractive module in the encoder-decoder model. However, these approaches, while effective in determining when to halt decoding, fall short in selecting relevant information to include within the specified length constraint. This article diverges from prior models reliant on predefined lengths. Instead, it introduces a novel approach to length-controllable abstractive summarization by integrating an image processing phase. This phase determines the specific size of the summary output slot. The proposed model harnesses enhanced T5 and GPT models, seamlessly adapting summaries to designated slots. The computed area of a given slot is employed in both models to generate abstractive summaries tailored to fit the output slot perfectly. Experimental evaluations on the CNN/Daily Mail dataset demonstrate the model's success in performing length-controlled summarization, yielding superior results.

摘要

摘要生成模型的最新进展,尤其是那些基于编码器-解码器架构构建的模型,通常会为每个源文本生成一个单一的摘要。控制摘要的长度对于实际应用至关重要,例如为不同篇幅大小的报纸或杂志撰写封面摘要。当前关于长度可控的摘要生成的研究采用了解码器模块中的长度嵌入或编码器-解码器模型中的词级提取模块等技术。然而,这些方法虽然在确定何时停止解码方面有效,但在选择要包含在指定长度约束内的相关信息方面存在不足。本文与依赖预定义长度的先前模型不同。相反,它通过集成图像处理阶段引入了一种新的长度可控摘要生成方法。这个阶段确定摘要输出槽的特定大小。所提出的模型利用增强的T5和GPT模型,无缝地使摘要适应指定的槽。给定槽的计算面积在两个模型中都被用于生成完美适合输出槽的摘要。在CNN/每日邮报数据集上的实验评估证明了该模型在执行长度控制摘要生成方面的成功,产生了优异的结果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/e83445374272/peerj-cs-11-2667-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/1bc2f1cf11b8/peerj-cs-11-2667-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/c4c2c4cc0396/peerj-cs-11-2667-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/be2df32964d3/peerj-cs-11-2667-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/7a0b378a6dfc/peerj-cs-11-2667-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/3868ae653a84/peerj-cs-11-2667-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/229b2c562cc7/peerj-cs-11-2667-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/d4f9115f22a2/peerj-cs-11-2667-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/583ac55427c8/peerj-cs-11-2667-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/793cb81b5ce7/peerj-cs-11-2667-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/8f79d3592d74/peerj-cs-11-2667-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/e83445374272/peerj-cs-11-2667-g011.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/1bc2f1cf11b8/peerj-cs-11-2667-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/c4c2c4cc0396/peerj-cs-11-2667-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/be2df32964d3/peerj-cs-11-2667-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/7a0b378a6dfc/peerj-cs-11-2667-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/3868ae653a84/peerj-cs-11-2667-g005.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/229b2c562cc7/peerj-cs-11-2667-g006.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/d4f9115f22a2/peerj-cs-11-2667-g007.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/583ac55427c8/peerj-cs-11-2667-g008.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/793cb81b5ce7/peerj-cs-11-2667-g009.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/8f79d3592d74/peerj-cs-11-2667-g010.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/90f1/11935775/e83445374272/peerj-cs-11-2667-g011.jpg

相似文献

1
Enhanced transformer for length-controlled abstractive summarization based on summary output area.基于摘要输出区域的用于长度控制的抽象摘要的增强型变压器。
PeerJ Comput Sci. 2025 Mar 11;11:e2667. doi: 10.7717/peerj-cs.2667. eCollection 2025.
2
Abstractive text summarization of low-resourced languages using deep learning.使用深度学习对低资源语言进行摘要性文本总结。
PeerJ Comput Sci. 2023 Jan 13;9:e1176. doi: 10.7717/peerj-cs.1176. eCollection 2023.
3
Unified extractive-abstractive summarization: a hybrid approach utilizing BERT and transformer models for enhanced document summarization.统一提取-抽象摘要:一种利用BERT和Transformer模型增强文档摘要的混合方法。
PeerJ Comput Sci. 2024 Nov 18;10:e2424. doi: 10.7717/peerj-cs.2424. eCollection 2024.
4
Multimodal Abstractive Summarization using bidirectional encoder representations from transformers with attention mechanism.使用带有注意力机制的变换器双向编码器表示的多模态抽象摘要
Heliyon. 2024 Feb 18;10(4):e26162. doi: 10.1016/j.heliyon.2024.e26162. eCollection 2024 Feb 29.
5
Dual Encoding for Abstractive Text Summarization.双重编码用于抽象文本摘要。
IEEE Trans Cybern. 2020 Mar;50(3):985-996. doi: 10.1109/TCYB.2018.2876317. Epub 2018 Nov 2.
6
Abstractive Arabic Text Summarization Based on Deep Learning.基于深度学习的摘要式阿拉伯文文本总结。
Comput Intell Neurosci. 2022 Jan 11;2022:1566890. doi: 10.1155/2022/1566890. eCollection 2022.
7
A data package for abstractive opinion summarization, title generation, and rating-based sentiment prediction for airline reviews.一个用于航空公司评论的抽象意见总结、标题生成和基于评分的情感预测的数据包。
Data Brief. 2023 Sep 1;50:109535. doi: 10.1016/j.dib.2023.109535. eCollection 2023 Oct.
8
Clinical research text summarization method based on fusion of domain knowledge.基于领域知识融合的临床研究文本摘要方法。
J Biomed Inform. 2024 Aug;156:104668. doi: 10.1016/j.jbi.2024.104668. Epub 2024 Jun 8.
9
Graph-based abstractive biomedical text summarization.基于图的抽象生物医学文本摘要
J Biomed Inform. 2022 Aug;132:104099. doi: 10.1016/j.jbi.2022.104099. Epub 2022 Jun 11.
10
Ontology-based prompt tuning for news article summarization.基于本体的新闻文章摘要提示调整
Front Artif Intell. 2025 Feb 11;8:1520144. doi: 10.3389/frai.2025.1520144. eCollection 2025.

本文引用的文献

1
Extractive text summarization system to aid data extraction from full text in systematic review development.用于从系统综述开发的全文中辅助数据提取的抽取式文本摘要系统。
J Biomed Inform. 2016 Dec;64:265-272. doi: 10.1016/j.jbi.2016.10.014. Epub 2016 Oct 27.