• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

堆叠型 DeBERTa:文本分类中针对不完整数据的全注意力

Stacked DeBERT: All attention in incomplete data for text classification.

机构信息

School Electronics and Electrical Engineering, South Korea.

Department of Artificial Intelligence, Kyungpook National University, Daegu, 41566, South Korea.

出版信息

Neural Netw. 2021 Apr;136:87-96. doi: 10.1016/j.neunet.2020.12.018. Epub 2020 Dec 25.

DOI:10.1016/j.neunet.2020.12.018
PMID:33453522
Abstract

In this paper, we propose Stacked DeBERT, short for StackedDenoising Bidirectional Encoder Representations from Transformers. This novel model improves robustness in incomplete data, when compared to existing systems, by designing a novel encoding scheme in BERT, a powerful language representation model solely based on attention mechanisms. Incomplete data in natural language processing refer to text with missing or incorrect words, and its presence can hinder the performance of current models that were not implemented to withstand such noises, but must still perform well even under duress. This is due to the fact that current approaches are built for and trained with clean and complete data, and thus are not able to extract features that can adequately represent incomplete data. Our proposed approach consists of obtaining intermediate input representations by applying an embedding layer to the input tokens followed by vanilla transformers. These intermediate features are given as input to novel denoising transformers which are responsible for obtaining richer input representations. The proposed approach takes advantage of stacks of multilayer perceptrons for the reconstruction of missing words' embeddings by extracting more abstract and meaningful hidden feature vectors, and bidirectional transformers for improved embedding representation. We consider two datasets for training and evaluation: the Chatbot Natural Language Understanding Evaluation Corpus and Kaggle's Twitter Sentiment Corpus. Our model shows improved F1-scores and better robustness in informal/incorrect texts present in tweets and in texts with Speech-to-Text error in the sentiment and intent classification tasks..

摘要

在本文中,我们提出了 Stacked DeBERT,它是 StackedDenoising Bidirectional Encoder Representations from Transformers 的缩写。与现有系统相比,该新型模型通过在 BERT 中设计一种新的编码方案,提高了在不完整数据下的鲁棒性。BERT 是一种强大的语言表示模型,仅基于注意力机制。自然语言处理中的不完整数据是指带有缺失或错误单词的文本,其存在会阻碍当前模型的性能,这些模型不是为了承受这种噪声而设计的,但即使在压力下也必须表现良好。这是因为当前的方法是为干净和完整的数据而构建和训练的,因此无法提取能够充分表示不完整数据的特征。我们提出的方法包括通过将嵌入层应用于输入令牌来获得中间输入表示,然后是香草变压器。这些中间特征作为输入提供给新的去噪变压器,负责获得更丰富的输入表示。所提出的方法利用多层感知机的堆栈通过提取更抽象和有意义的隐藏特征向量来重建缺失单词的嵌入,以及双向变压器来改善嵌入表示。我们考虑了两个用于训练和评估的数据集:Chatbot Natural Language Understanding Evaluation Corpus 和 Kaggle 的 Twitter Sentiment Corpus。我们的模型在情感和意图分类任务中,在带有 Speech-to-Text 错误的文本和带有非正式/错误的推文的文本中,显示出了更高的 F1 分数和更好的鲁棒性。

相似文献

1
Stacked DeBERT: All attention in incomplete data for text classification.堆叠型 DeBERTa:文本分类中针对不完整数据的全注意力
Neural Netw. 2021 Apr;136:87-96. doi: 10.1016/j.neunet.2020.12.018. Epub 2020 Dec 25.
2
Limitations of Transformers on Clinical Text Classification.Transformer 在临床文本分类上的局限性。
IEEE J Biomed Health Inform. 2021 Sep;25(9):3596-3607. doi: 10.1109/JBHI.2021.3062322. Epub 2021 Sep 3.
3
GT-Finder: Classify the family of glucose transporters with pre-trained BERT language models.GT-Finder:使用预训练的 BERT 语言模型对葡萄糖转运蛋白家族进行分类。
Comput Biol Med. 2021 Apr;131:104259. doi: 10.1016/j.compbiomed.2021.104259. Epub 2021 Feb 7.
4
Oversampling effect in pretraining for bidirectional encoder representations from transformers (BERT) to localize medical BERT and enhance biomedical BERT.在基于转换器的双向编码器表示预训练(BERT)中进行过采样,以定位医学 BERT 并增强生物医学 BERT。
Artif Intell Med. 2024 Jul;153:102889. doi: 10.1016/j.artmed.2024.102889. Epub 2024 May 5.
5
Text Sentiment Classification Based on BERT Embedding and Sliced Multi-Head Self-Attention Bi-GRU.基于 BERT 嵌入和切片多头自注意力 Bi-GRU 的文本情感分类
Sensors (Basel). 2023 Jan 28;23(3):1481. doi: 10.3390/s23031481.
6
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.
7
Automatic text classification of actionable radiology reports of tinnitus patients using bidirectional encoder representations from transformer (BERT) and in-domain pre-training (IDPT).使用基于转换器的双向编码器表示 (BERT) 和领域内预训练 (IDPT) 对耳鸣患者的可操作放射学报告进行自动文本分类。
BMC Med Inform Decis Mak. 2022 Jul 30;22(1):200. doi: 10.1186/s12911-022-01946-y.
8
Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.基于RoBERTa-WWM-ext + CNN(带有全词掩码扩展的基于变换器预训练方法的稳健优化双向编码器表示与卷积神经网络相结合)模型的医患对话多标签分类:命名实体研究
JMIR Med Inform. 2022 Apr 21;10(4):e35606. doi: 10.2196/35606.
9
A BERT based dual-channel explainable text emotion recognition system.基于 BERT 的双通道可解释文本情感识别系统。
Neural Netw. 2022 Jun;150:392-407. doi: 10.1016/j.neunet.2022.03.017. Epub 2022 Mar 18.
10
Interactive Dual Attention Network for Text Sentiment Classification.用于文本情感分类的交互式双注意力网络。
Comput Intell Neurosci. 2020 Nov 3;2020:8858717. doi: 10.1155/2020/8858717. eCollection 2020.