• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

基于Transformer的自然语言处理综述。

Natural language processing with transformers: a review.

作者信息

Tucudean Georgiana, Bucos Marian, Dragulescu Bogdan, Caleanu Catalin Daniel

机构信息

Communications Department, Politehnica University Timișoara, Timișoara, Timiș, România.

Applied Electronics Department, Politehnica University Timișoara, Timișoara, Timiș, România.

出版信息

PeerJ Comput Sci. 2024 Aug 7;10:e2222. doi: 10.7717/peerj-cs.2222. eCollection 2024.

DOI:10.7717/peerj-cs.2222
PMID:39145251
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC11322986/
Abstract

Natural language processing (NLP) tasks can be addressed with several deep learning architectures, and many different approaches have proven to be efficient. This study aims to briefly summarize the use cases for NLP tasks along with the main architectures. This research presents transformer-based solutions for NLP tasks such as Bidirectional Encoder Representations from Transformers (BERT), and Generative Pre-Training (GPT) architectures. To achieve that, we conducted a step-by-step process in the review strategy: identify the recent studies that include Transformers, apply filters to extract the most consistent studies, identify and define inclusion and exclusion criteria, assess the strategy proposed in each study, and finally discuss the methods and architectures presented in the resulting articles. These steps facilitated the systematic summarization and comparative analysis of NLP applications based on Transformer architectures. The primary focus is the current state of the NLP domain, particularly regarding its applications, language models, and data set types. The results provide insights into the challenges encountered in this research domain.

摘要

自然语言处理(NLP)任务可以通过多种深度学习架构来解决,并且许多不同的方法已被证明是有效的。本研究旨在简要总结NLP任务的用例以及主要架构。本研究提出了用于NLP任务的基于Transformer的解决方案,如来自Transformer的双向编码器表示(BERT)和生成式预训练(GPT)架构。为了实现这一目标,我们在综述策略中进行了一个循序渐进的过程:识别最近包含Transformer的研究,应用过滤器提取最一致的研究,识别和定义纳入和排除标准,评估每项研究中提出的策略,最后讨论所得文章中提出的方法和架构。这些步骤有助于基于Transformer架构对NLP应用进行系统的总结和比较分析。主要重点是NLP领域的当前状态,特别是关于其应用、语言模型和数据集类型。研究结果为该研究领域中遇到的挑战提供了见解。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c24b/11322986/c14857a0aeba/peerj-cs-10-2222-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c24b/11322986/c14857a0aeba/peerj-cs-10-2222-g001.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/c24b/11322986/c14857a0aeba/peerj-cs-10-2222-g001.jpg

相似文献

1
Natural language processing with transformers: a review.基于Transformer的自然语言处理综述。
PeerJ Comput Sci. 2024 Aug 7;10:e2222. doi: 10.7717/peerj-cs.2222. eCollection 2024.
2
Few-Shot Learning for Clinical Natural Language Processing Using Siamese Neural Networks: Algorithm Development and Validation Study.使用暹罗神经网络的临床自然语言处理少样本学习:算法开发与验证研究
JMIR AI. 2023 May 4;2:e44293. doi: 10.2196/44293.
3
Reading comprehension based question answering system in Bangla language with transformer-based learning.基于基于变压器学习的孟加拉语阅读理解问答系统。
Heliyon. 2022 Oct 12;8(10):e11052. doi: 10.1016/j.heliyon.2022.e11052. eCollection 2022 Oct.
4
Comparing deep learning architectures for sentiment analysis on drug reviews.比较药物评论情感分析的深度学习架构。
J Biomed Inform. 2020 Oct;110:103539. doi: 10.1016/j.jbi.2020.103539. Epub 2020 Aug 17.
5
Transformers-sklearn: a toolkit for medical language understanding with transformer-based models.Transformer-sklearn:一个基于 Transformer 的模型的医学语言理解工具包。
BMC Med Inform Decis Mak. 2021 Jul 30;21(Suppl 2):90. doi: 10.1186/s12911-021-01459-0.
6
Bidirectional Encoder Representations from Transformers in Radiology: A Systematic Review of Natural Language Processing Applications.基于 Transformer 的双向编码器表示在放射学中的应用:自然语言处理应用的系统评价。
J Am Coll Radiol. 2024 Jun;21(6):914-941. doi: 10.1016/j.jacr.2024.01.012. Epub 2024 Jan 30.
7
A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.从自由文本肿瘤病理学报告(CancerBERT 网络)中提取数据的问答系统:开发研究。
J Med Internet Res. 2022 Mar 23;24(3):e27210. doi: 10.2196/27210.
8
RadBERT: Adapting Transformer-based Language Models to Radiology.RadBERT:使基于Transformer的语言模型适用于放射学领域。
Radiol Artif Intell. 2022 Jun 15;4(4):e210258. doi: 10.1148/ryai.210258. eCollection 2022 Jul.
9
Measurement of Semantic Textual Similarity in Clinical Texts: Comparison of Transformer-Based Models.临床文本中语义文本相似度的测量:基于Transformer模型的比较。
JMIR Med Inform. 2020 Nov 23;8(11):e19735. doi: 10.2196/19735.
10
Applications of Natural Language Processing for the Management of Stroke Disorders: Scoping Review.自然语言处理在中风疾病管理中的应用:范围综述
JMIR Med Inform. 2023 Sep 6;11:e48693. doi: 10.2196/48693.

引用本文的文献

1
deep-Sep: a deep learning-based method for fast and accurate prediction of selenoprotein genes in bacteria.Deep-Sep:一种基于深度学习的快速准确预测细菌中硒蛋白基因的方法。
mSystems. 2025 Apr 22;10(4):e0125824. doi: 10.1128/msystems.01258-24. Epub 2025 Mar 10.

本文引用的文献

1
Natural language processing: state of the art, current trends and challenges.自然语言处理:技术现状、当前趋势与挑战。
Multimed Tools Appl. 2023;82(3):3713-3744. doi: 10.1007/s11042-022-13428-4. Epub 2022 Jul 14.
2
MolGPT: Molecular Generation Using a Transformer-Decoder Model.MolGPT:基于 Transformer-Decoder 模型的分子生成。
J Chem Inf Model. 2022 May 9;62(9):2064-2076. doi: 10.1021/acs.jcim.1c00600. Epub 2021 Oct 25.
3
Med-BERT: pretrained contextualized embeddings on large-scale structured electronic health records for disease prediction.
医学BERT:基于大规模结构化电子健康记录进行疾病预测的预训练上下文嵌入模型
NPJ Digit Med. 2021 May 20;4(1):86. doi: 10.1038/s41746-021-00455-y.
4
Combat COVID-19 infodemic using explainable natural language processing models.使用可解释的自然语言处理模型应对新冠疫情信息疫情。
Inf Process Manag. 2021 Jul;58(4):102569. doi: 10.1016/j.ipm.2021.102569. Epub 2021 Mar 6.
5
Molecular optimization by capturing chemist's intuition using deep neural networks.通过使用深度神经网络捕捉化学家的直觉进行分子优化。
J Cheminform. 2021 Mar 20;13(1):26. doi: 10.1186/s13321-021-00497-0.
6
Limitations of Transformers on Clinical Text Classification.Transformer 在临床文本分类上的局限性。
IEEE J Biomed Health Inform. 2021 Sep;25(9):3596-3607. doi: 10.1109/JBHI.2021.3062322. Epub 2021 Sep 3.
7
A transformer architecture based on BERT and 2D convolutional neural network to identify DNA enhancers from sequence information.基于 BERT 和二维卷积神经网络的变压器架构,用于从序列信息中识别 DNA 增强子。
Brief Bioinform. 2021 Sep 2;22(5). doi: 10.1093/bib/bbab005.
8
Clinical concept extraction using transformers.使用转换器进行临床概念提取。
J Am Med Inform Assoc. 2020 Dec 9;27(12):1935-1942. doi: 10.1093/jamia/ocaa189.
9
Comparing deep learning architectures for sentiment analysis on drug reviews.比较药物评论情感分析的深度学习架构。
J Biomed Inform. 2020 Oct;110:103539. doi: 10.1016/j.jbi.2020.103539. Epub 2020 Aug 17.
10
Extracting comprehensive clinical information for breast cancer using deep learning methods.利用深度学习方法提取乳腺癌全面临床信息。
Int J Med Inform. 2019 Dec;132:103985. doi: 10.1016/j.ijmedinf.2019.103985. Epub 2019 Oct 2.