Suppr超能文献

N-GPETS:基于神经注意力图的抽取式文本摘要预训练统计模型。

N-GPETS: Neural Attention Graph-Based Pretrained Statistical Model for Extractive Text Summarization.

机构信息

Department of Computer Science, City University of Science and Information Technology, Peshawar 25000, Pakistan.

Department of Computer Science, Islamia College, Peshawar 25000, Pakistan.

出版信息

Comput Intell Neurosci. 2022 Nov 22;2022:6241373. doi: 10.1155/2022/6241373. eCollection 2022.

Abstract

The extractive summarization approach involves selecting the source document's salient sentences to build a summary. One of the most important aspects of extractive summarization is learning and modelling cross-sentence associations. Inspired by the popularity of Transformer-based Bidirectional Encoder Representations (BERT) pretrained linguistic model and graph attention network (GAT) having a sophisticated network that captures intersentence associations, this research work proposes a novel neural model N-GPETS by combining heterogeneous graph attention network with BERT model along with statistical approach using TF-IDF values for extractive summarization task. Apart from sentence nodes, N-GPETS also works with different semantic word nodes of varying granularity levels that serve as a link between sentences, improving intersentence interaction. Furthermore, proposed N-GPETS becomes more improved and feature-rich by integrating graph layer with BERT encoder at graph initialization step rather than employing other neural network encoders such as CNN or LSTM. To the best of our knowledge, this work is the first attempt to combine the BERT encoder and TF-IDF values of the entire document with a heterogeneous attention graph structure for the extractive summarization task. The empirical outcomes on benchmark news data sets CNN/DM show that the proposed model N-GPETS gets favorable results in comparison with other heterogeneous graph structures employing the BERT model and graph structures without the BERT model.

摘要

抽取式摘要方法涉及选择源文档的突出句子来构建摘要。抽取式摘要的最重要方面之一是学习和建模跨句子关联。受基于 Transformer 的双向编码器表示(BERT)预训练语言模型和图注意网络(GAT)的普及的启发,GAT 具有捕捉句子间关联的复杂网络,本研究工作通过结合异构图注意网络和 BERT 模型以及使用 TF-IDF 值的统计方法,提出了一种新颖的神经模型 N-GPETS,用于抽取式摘要任务。除了句子节点外,N-GPETS 还使用不同粒度级别的不同语义词节点作为句子之间的链接,从而提高句子间的交互作用。此外,通过在图初始化步骤中与 BERT 编码器集成图层,而不是使用其他神经网络编码器(如 CNN 或 LSTM),提出的 N-GPETS 变得更加改进和丰富。据我们所知,这项工作首次尝试将 BERT 编码器和整个文档的 TF-IDF 值与异构图注意力结构相结合,用于抽取式摘要任务。在基准新闻数据集 CNN/DM 上的实验结果表明,与其他使用 BERT 模型和无 BERT 模型的异构图结构的模型相比,所提出的模型 N-GPETS 具有更好的效果。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/9670/9708337/ae7d044e898b/CIN2022-6241373.001.jpg

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验