Su Wu, Jiang Jin, Huang Kaihui
School of Automation and Electronic Information, Xiangtan University, Xiangtan, Hunan Province, China.
PeerJ Comput Sci. 2023 Dec 13;9:e1737. doi: 10.7717/peerj-cs.1737. eCollection 2023.
The crucial aspect of extractive document summarization lies in understanding the interrelations between sentences. Documents inherently comprise a multitude of sentences, and sentence-level models frequently fail to consider the relationships between distantly-placed sentences, resulting in the omission of significant information in the summary. Moreover, information within documents tends to be distributed sparsely, challenging the efficacy of sentence-level models. In the realm of heterogeneous graph neural networks, it has been observed that semantic nodes with varying levels of granularity encapsulate distinct semantic connections. Initially, the incorporation of edge features into the computation of dynamic graph attention networks is performed to account for node relationships. Subsequently, given the multiplicity of topics in a document or a set of documents, a topic model is employed to extract topic-specific features and the probability distribution linking these topics with sentence nodes. Last but not least, the model defines nodes with different levels of granularity-ranging from documents and topics to sentences-and these various nodes necessitate different propagation widths and depths for capturing intricate relationships in the information being disseminated. Adaptive measures are taken to learn the importance and correlation between nodes of different granularities in terms of both width and depth. Experimental evidence from two benchmark datasets highlights the superior performance of the proposed model, as assessed by ROUGE metrics, in comparison to existing approaches, even in the absence of pre-trained language models. Additionally, an ablation study confirms the positive impact of each individual module on the model's ROUGE scores.
提取式文档摘要的关键在于理解句子之间的相互关系。文档本质上由大量句子组成,而句子级模型常常无法考虑距离较远的句子之间的关系,导致摘要中遗漏重要信息。此外,文档中的信息往往分布稀疏,这对句子级模型的有效性构成挑战。在异构图神经网络领域,人们观察到具有不同粒度级别的语义节点封装了不同的语义连接。首先,将边特征纳入动态图注意力网络的计算中,以考虑节点关系。随后,鉴于文档或一组文档中主题的多样性,采用主题模型来提取特定主题的特征以及将这些主题与句子节点联系起来的概率分布。最后但同样重要的是,该模型定义了具有不同粒度级别的节点——从文档、主题到句子——而这些不同的节点需要不同的传播宽度和深度来捕捉正在传播的信息中的复杂关系。采取自适应措施来学习不同粒度的节点在宽度和深度方面的重要性和相关性。来自两个基准数据集的实验证据表明,与现有方法相比,即使在没有预训练语言模型的情况下,通过ROUGE指标评估,所提出的模型也具有卓越的性能。此外,一项消融研究证实了每个单独模块对模型ROUGE分数的积极影响。