• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Limitations of Transformers on Clinical Text Classification.Transformer 在临床文本分类上的局限性。
IEEE J Biomed Health Inform. 2021 Sep;25(9):3596-3607. doi: 10.1109/JBHI.2021.3062322. Epub 2021 Sep 3.
2
When BERT meets Bilbo: a learning curve analysis of pretrained language model on disease classification.当 BERT 遇见比尔博:预训练语言模型在疾病分类上的学习曲线分析。
BMC Med Inform Decis Mak. 2022 Apr 5;21(Suppl 9):377. doi: 10.1186/s12911-022-01829-2.
3
Oversampling effect in pretraining for bidirectional encoder representations from transformers (BERT) to localize medical BERT and enhance biomedical BERT.在基于转换器的双向编码器表示预训练(BERT)中进行过采样,以定位医学 BERT 并增强生物医学 BERT。
Artif Intell Med. 2024 Jul;153:102889. doi: 10.1016/j.artmed.2024.102889. Epub 2024 May 5.
4
Stacked DeBERT: All attention in incomplete data for text classification.堆叠型 DeBERTa:文本分类中针对不完整数据的全注意力
Neural Netw. 2021 Apr;136:87-96. doi: 10.1016/j.neunet.2020.12.018. Epub 2020 Dec 25.
5
Classifying social determinants of health from unstructured electronic health records using deep learning-based natural language processing.利用基于深度学习的自然语言处理技术从非结构化电子健康记录中分类社会健康决定因素。
J Biomed Inform. 2022 Mar;127:103984. doi: 10.1016/j.jbi.2021.103984. Epub 2022 Jan 7.
6
Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.基于RoBERTa-WWM-ext + CNN(带有全词掩码扩展的基于变换器预训练方法的稳健优化双向编码器表示与卷积神经网络相结合)模型的医患对话多标签分类:命名实体研究
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
7
Chinese text classification by combining Chinese-BERTology-wwm and GCN.结合中文BERTology-wwm和图卷积网络进行中文文本分类。
PeerJ Comput Sci. 2023 Aug 17;9:e1544. doi: 10.7717/peerj-cs.1544. eCollection 2023.
8
Use of BERT (Bidirectional Encoder Representations from Transformers)-Based Deep Learning Method for Extracting Evidences in Chinese Radiology Reports: Development of a Computer-Aided Liver Cancer Diagnosis Framework.基于 BERT(来自 Transformers 的双向编码器表示)的深度学习方法在提取中文放射学报告证据中的应用:计算机辅助肝癌诊断框架的开发。
J Med Internet Res. 2021 Jan 12;23(1):e19689. doi: 10.2196/19689.
9
A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.从自由文本肿瘤病理学报告(CancerBERT 网络)中提取数据的问答系统:开发研究。
J Med Internet Res. 2022 Mar 23;24(3):e27210. doi: 10.2196/27210.
10
Korean clinical entity recognition from diagnosis text using BERT.基于 BERT 的韩语文本临床实体识别。
BMC Med Inform Decis Mak. 2020 Sep 30;20(Suppl 7):242. doi: 10.1186/s12911-020-01241-8.

引用本文的文献

1
Leveraging BERT for embedding ICD codes from large scale cardiovascular EMR data to understand patient diagnostic patterns.利用BERT从大规模心血管电子病历数据中嵌入ICD编码,以了解患者的诊断模式。
BMC Med Inform Decis Mak. 2025 Aug 11;25(1):300. doi: 10.1186/s12911-025-03145-x.
2
Improving Transformer Performance for French Clinical Notes Classification Using Mixture of Experts on a Limited Dataset.在有限数据集上使用专家混合策略提高用于法语临床笔记分类的Transformer性能
IEEE J Transl Eng Health Med. 2025 Jun 4;13:261-274. doi: 10.1109/JTEHM.2025.3576570. eCollection 2025.
3
A survey of NLP methods for oncology in the past decade with a focus on cancer registry applications.对过去十年肿瘤学领域自然语言处理方法的一项调查,重点关注癌症登记应用。
Artif Intell Rev. 2025;58(10):314. doi: 10.1007/s10462-025-11316-5. Epub 2025 Jul 16.
4
Language Models for Multilabel Document Classification of Surgical Concepts in Exploratory Laparotomy Operative Notes: Algorithm Development Study.用于探索性剖腹手术记录中手术概念多标签文档分类的语言模型:算法开发研究
JMIR Med Inform. 2025 Jul 9;13:e71176. doi: 10.2196/71176.
5
Improving Emergency Department Visit Risk Prediction: Exploring the Operational Utility of Applied Patient Portal Messages.改善急诊科就诊风险预测:探索应用患者门户网站信息的操作效用。
AMIA Annu Symp Proc. 2025 May 22;2024:610-619. eCollection 2024.
6
Enhanced effective convolutional attention network with squeeze-and-excitation inception module for multi-label clinical document classification.基于挤压激励 inception 模块的增强型有效卷积注意力网络用于多标签临床文档分类
Sci Rep. 2025 May 16;15(1):16988. doi: 10.1038/s41598-025-98719-0.
7
MISTIC: a novel approach for metastasis classification in Italian electronic health records using transformers.MISTIC:一种使用变压器对意大利电子健康记录中的转移进行分类的新方法。
BMC Med Inform Decis Mak. 2025 Apr 10;25(1):160. doi: 10.1186/s12911-025-02994-w.
8
Large Language Model Applications for Health Information Extraction in Oncology: Scoping Review.用于肿瘤学健康信息提取的大语言模型应用:范围综述
JMIR Cancer. 2025 Mar 28;11:e65984. doi: 10.2196/65984.
9
Data transformation of unstructured electroencephalography reports by natural language processing: improving data usability for large-scale epilepsy studies.通过自然语言处理对非结构化脑电图报告进行数据转换:提高大规模癫痫研究的数据可用性。
Front Neurol. 2025 Feb 28;16:1521001. doi: 10.3389/fneur.2025.1521001. eCollection 2025.
10
Microblog discourse analysis for parenting style assessment.用于育儿风格评估的微博话语分析
Front Public Health. 2025 Feb 11;13:1505825. doi: 10.3389/fpubh.2025.1505825. eCollection 2025.

本文引用的文献

1
Classification of Cancer Pathology Reports: A Large-Scale Comparative Study.癌症病理学报告分类:一项大规模比较研究。
IEEE J Biomed Health Inform. 2020 Nov;24(11):3085-3094. doi: 10.1109/JBHI.2020.3005016. Epub 2020 Nov 4.
2
Deep learning for electronic health records: A comparative review of multiple deep neural architectures.深度学习在电子健康记录中的应用:多种深度神经网络架构的比较综述。
J Biomed Inform. 2020 Jan;101:103337. doi: 10.1016/j.jbi.2019.103337.
3
Classifying cancer pathology reports with hierarchical self-attention networks.基于层次自注意力网络的癌症病理报告分类。
Artif Intell Med. 2019 Nov;101:101726. doi: 10.1016/j.artmed.2019.101726. Epub 2019 Oct 15.
4
Deep learning in clinical natural language processing: a methodical review.深度学习在临床自然语言处理中的应用:系统综述。
J Am Med Inform Assoc. 2020 Mar 1;27(3):457-470. doi: 10.1093/jamia/ocz200.
5
Automatic extraction of cancer registry reportable information from free-text pathology reports using multitask convolutional neural networks.使用多任务卷积神经网络从自由文本病理报告中自动提取癌症登记报告信息。
J Am Med Inform Assoc. 2020 Jan 1;27(1):89-98. doi: 10.1093/jamia/ocz153.
6
BioBERT: a pre-trained biomedical language representation model for biomedical text mining.BioBERT:一种用于生物医学文本挖掘的预训练生物医学语言表示模型。
Bioinformatics. 2020 Feb 15;36(4):1234-1240. doi: 10.1093/bioinformatics/btz682.
7
BioWordVec, improving biomedical word embeddings with subword information and MeSH.BioWordVec,利用子词信息和 MeSH 改进生物医学词向量。
Sci Data. 2019 May 10;6(1):52. doi: 10.1038/s41597-019-0055-0.
8
Natural Language Processing of Clinical Notes on Chronic Diseases: Systematic Review.慢性病临床记录的自然语言处理:系统综述
JMIR Med Inform. 2019 Apr 27;7(2):e12239. doi: 10.2196/12239.
9
Opportunities and challenges in developing deep learning models using electronic health records data: a systematic review.利用电子健康记录数据开发深度学习模型的机遇与挑战:系统综述。
J Am Med Inform Assoc. 2018 Oct 1;25(10):1419-1428. doi: 10.1093/jamia/ocy068.
10
Automated extraction of Biomarker information from pathology reports.从病理报告中自动提取生物标志物信息。
BMC Med Inform Decis Mak. 2018 May 21;18(1):29. doi: 10.1186/s12911-018-0609-7.

Transformer 在临床文本分类上的局限性。

Limitations of Transformers on Clinical Text Classification.

出版信息

IEEE J Biomed Health Inform. 2021 Sep;25(9):3596-3607. doi: 10.1109/JBHI.2021.3062322. Epub 2021 Sep 3.

DOI:10.1109/JBHI.2021.3062322
PMID:33635801
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC8387496/
Abstract

Bidirectional Encoder Representations from Transformers (BERT) and BERT-based approaches are the current state-of-the-art in many natural language processing (NLP) tasks; however, their application to document classification on long clinical texts is limited. In this work, we introduce four methods to scale BERT, which by default can only handle input sequences up to approximately 400 words long, to perform document classification on clinical texts several thousand words long. We compare these methods against two much simpler architectures - a word-level convolutional neural network and a hierarchical self-attention network - and show that BERT often cannot beat these simpler baselines when classifying MIMIC-III discharge summaries and SEER cancer pathology reports. In our analysis, we show that two key components of BERT - pretraining and WordPiece tokenization - may actually be inhibiting BERT's performance on clinical text classification tasks where the input document is several thousand words long and where correctly identifying labels may depend more on identifying a few key words or phrases rather than understanding the contextual meaning of sequences of text.

摘要

基于转换器的双向编码器表示 (BERT) 和基于 BERT 的方法是许多自然语言处理 (NLP) 任务的当前最新技术; 然而,它们在长临床文本的文档分类中的应用受到限制。在这项工作中,我们引入了四种扩展 BERT 的方法,BERT 默认只能处理长度约为 400 字的输入序列,以对数千字长的临床文本进行文档分类。我们将这些方法与两个简单得多的架构进行了比较——一个是单词级别的卷积神经网络,另一个是分层自注意网络——并表明,在对 MIMIC-III 出院总结和 SEER 癌症病理报告进行分类时,BERT 通常无法击败这些更简单的基线。在我们的分析中,我们表明 BERT 的两个关键组成部分——预训练和 WordPiece 标记化——实际上可能会抑制 BERT 在临床文本分类任务中的性能,在这些任务中,输入文档长达数千字,正确识别标签可能更多地取决于识别几个关键词或短语,而不是理解文本序列的上下文含义。