• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于段落级注意力的深度模型用于章节分割。

Paragraph-level attention based deep model for chapter segmentation.

作者信息

Virameteekul Paveen

机构信息

Department of Computer Science & Engineering, Shanghai Jiao Tong University, Minhang, Shanghai, China.

出版信息

PeerJ Comput Sci. 2022 Jun 10;8:e1003. doi: 10.7717/peerj-cs.1003. eCollection 2022.

DOI:10.7717/peerj-cs.1003
PMID:35721402
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC9202623/
Abstract

Books are usually divided into chapters and sections. Correctly and automatically recognizing chapter boundaries can work as a proxy when segmenting long texts (a more general task). Book chapters can be easily segmented by humans, but automatic segregation is more challenging because the data is semi-structured. Since the concept of language is prone to ambiguity, it is essential to identify the relationship between the words in each paragraph and classify each consecutive paragraph based on their respective relationships with one another. Although researchers have designed deep learning-based models to solve this problem, these approaches have not considered the paragraph-level semantics among the consecutive paragraphs. In this article, we propose a novel deep learning-based method to segment book chapters that uses paragraph-level semantics and an attention mechanism. We first utilized a pre-trained XLNet model connected to a convolutional neural network (CNN) to extract the semantic meaning of each paragraph. Then, we measured the similarities in the semantics of each paragraph and designed an attention mechanism to inject the similarity information in order to better predict the chapter boundaries. The experimental results indicated that the performance of our proposed method can surpass those of other state-of-the-art (SOTA) methods for chapter segmentation on public datasets (the proposed model achieved an F1 score of 0.8856, outperforming the Bidirectional Encoder Representations from Transformers (BERT) model's F1 score of 0.6640). The ablation study also illustrated that the paragraph-level attention mechanism could produce a significant increase in performance.

摘要

书籍通常分为章节和小节。在分割长文本(一项更具普遍性的任务)时,正确且自动地识别章节边界可作为一种替代方法。书籍章节很容易被人类分割,但自动分割更具挑战性,因为数据是半结构化的。由于语言概念容易产生歧义,因此识别每个段落中单词之间的关系并根据它们彼此之间的各自关系对每个连续段落进行分类至关重要。尽管研究人员已经设计了基于深度学习的模型来解决这个问题,但这些方法没有考虑连续段落之间的段落级语义。在本文中,我们提出了一种基于深度学习的新颖方法来分割书籍章节,该方法使用段落级语义和注意力机制。我们首先利用一个预训练的XLNet模型连接到卷积神经网络(CNN)来提取每个段落的语义。然后,我们测量每个段落语义的相似度,并设计一种注意力机制来注入相似度信息,以便更好地预测章节边界。实验结果表明,我们提出的方法在公共数据集上的章节分割性能可以超过其他现有最先进(SOTA)方法(所提出的模型实现了0.8856的F1分数,优于来自Transformer的双向编码器表示(BERT)模型的0.6640的F1分数)。消融研究还表明,段落级注意力机制可以显著提高性能。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/8e0ba5f0c3a9/peerj-cs-08-1003-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/58a1bc6b18c1/peerj-cs-08-1003-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/4dd267c679bb/peerj-cs-08-1003-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/895c6fb19550/peerj-cs-08-1003-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/8e0ba5f0c3a9/peerj-cs-08-1003-g004.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/58a1bc6b18c1/peerj-cs-08-1003-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/4dd267c679bb/peerj-cs-08-1003-g002.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/895c6fb19550/peerj-cs-08-1003-g003.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/741f/9202623/8e0ba5f0c3a9/peerj-cs-08-1003-g004.jpg

相似文献

1
Paragraph-level attention based deep model for chapter segmentation.基于段落级注意力的深度模型用于章节分割。
PeerJ Comput Sci. 2022 Jun 10;8:e1003. doi: 10.7717/peerj-cs.1003. eCollection 2022.
2
Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.基于RoBERTa-WWM-ext + CNN(带有全词掩码扩展的基于变换器预训练方法的稳健优化双向编码器表示与卷积神经网络相结合)模型的医患对话多标签分类:命名实体研究
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
3
Predicting Semantic Similarity Between Clinical Sentence Pairs Using Transformer Models: Evaluation and Representational Analysis.使用Transformer模型预测临床句子对之间的语义相似性:评估与表征分析
JMIR Med Inform. 2021 May 26;9(5):e23099. doi: 10.2196/23099.
4
Fine-tuning BERT for automatic ADME semantic labeling in FDA drug labeling to enhance product-specific guidance assessment.在FDA药品标签中微调BERT以进行自动ADME语义标注,以加强特定产品的指导评估。
J Biomed Inform. 2023 Feb;138:104285. doi: 10.1016/j.jbi.2023.104285. Epub 2023 Jan 9.
5
A transformer architecture based on BERT and 2D convolutional neural network to identify DNA enhancers from sequence information.基于 BERT 和二维卷积神经网络的变压器架构,用于从序列信息中识别 DNA 增强子。
Brief Bioinform. 2021 Sep 2;22(5). doi: 10.1093/bib/bbab005.
6
Relation Classification for Bleeding Events From Electronic Health Records Using Deep Learning Systems: An Empirical Study.使用深度学习系统对电子健康记录中的出血事件进行关系分类:一项实证研究。
JMIR Med Inform. 2021 Jul 2;9(7):e27527. doi: 10.2196/27527.
7
A comparative study on deep learning models for text classification of unstructured medical notes with various levels of class imbalance.深度学习模型在不同类别不平衡程度的非结构化医疗记录文本分类中的对比研究。
BMC Med Res Methodol. 2022 Jul 2;22(1):181. doi: 10.1186/s12874-022-01665-y.
8
Multi-Level Representation Learning for Chinese Medical Entity Recognition: Model Development and Validation.用于中文医学实体识别的多层次表示学习:模型开发与验证
JMIR Med Inform. 2020 May 4;8(5):e17637. doi: 10.2196/17637.
9
Use of BERT (Bidirectional Encoder Representations from Transformers)-Based Deep Learning Method for Extracting Evidences in Chinese Radiology Reports: Development of a Computer-Aided Liver Cancer Diagnosis Framework.基于 BERT(来自 Transformers 的双向编码器表示)的深度学习方法在提取中文放射学报告证据中的应用:计算机辅助肝癌诊断框架的开发。
J Med Internet Res. 2021 Jan 12;23(1):e19689. doi: 10.2196/19689.
10
Deep Learning Approach for Negation and Speculation Detection for Automated Important Finding Flagging and Extraction in Radiology Report: Internal Validation and Technique Comparison Study.用于放射学报告中自动重要发现标记和提取的否定与推测检测的深度学习方法:内部验证与技术比较研究
JMIR Med Inform. 2023 Apr 25;11:e46348. doi: 10.2196/46348.

本文引用的文献

1
Domain Generalization for Named Entity Boundary Detection via Metalearning.基于元学习的命名实体边界检测的领域泛化。
IEEE Trans Neural Netw Learn Syst. 2021 Sep;32(9):3819-3830. doi: 10.1109/TNNLS.2020.3015912. Epub 2021 Aug 31.
2
A Comparative Analysis of Machine/Deep Learning Models for Parking Space Availability Prediction.基于机器/深度学习模型的停车位可用预测的比较分析。
Sensors (Basel). 2020 Jan 6;20(1):322. doi: 10.3390/s20010322.