• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

MCWS-Transformer:通过基于多上下文窗口的缩放自注意力实现蛋白质序列的高效建模

MCWS-Transformers: Towards an Efficient Modeling of Protein Sequences via Multi Context-Window Based Scaled Self-Attention.

作者信息

Ranjan Ashish, Fahad Md Shah, Fernandez-Baca David, Tripathi Sudhakar, Deepak Akshay

出版信息

IEEE/ACM Trans Comput Biol Bioinform. 2023 Mar-Apr;20(2):1188-1199. doi: 10.1109/TCBB.2022.3173789. Epub 2023 Apr 3.

DOI:10.1109/TCBB.2022.3173789
PMID:35536815
Abstract

This paper advances the self-attention mechanism in the standard transformer network specific to the modeling of the protein sequences. We introduce a novel context-window based scaled self-attention mechanism for processing protein sequences that is based on the notion of (i) local context and (ii) large contextual pattern. Both notions are essential to building a good representation for protein sequences. The proposed context-window based scaled self-attention mechanism is further used to build the multi context-window based scaled (MCWS) transformer network for the protein function prediction task at the protein sub-sequence level. Overall, the proposed MCWS transformer network produced improved predictive performances, outperforming existing state-of-the-art approaches by substantial margins. With respect to the standard transformer network, the proposed network produced improvements in F1-score of +2.30% and +2.08% on the biological process (BP) and molecular function (MF) datasets, respectively. The corresponding improvements over the state-of-the-art ProtVecGen-Plus+ProtVecGen-Ensemble approach are +3.38% (BP) and +2.86% (MF). Equally important, robust performances were obtained across protein sequences of different lengths.

摘要

本文改进了标准变压器网络中用于蛋白质序列建模的自注意力机制。我们引入了一种新颖的基于上下文窗口的缩放自注意力机制来处理蛋白质序列,该机制基于(i)局部上下文和(ii)大上下文模式的概念。这两个概念对于构建蛋白质序列的良好表示都至关重要。所提出的基于上下文窗口的缩放自注意力机制进一步用于构建基于多上下文窗口的缩放(MCWS)变压器网络,用于蛋白质子序列水平的蛋白质功能预测任务。总体而言,所提出的MCWS变压器网络产生了改进的预测性能,大幅优于现有的最先进方法。相对于标准变压器网络,所提出的网络在生物过程(BP)和分子功能(MF)数据集上的F1分数分别提高了+2.30%和+2.08%。相对于最先进的ProtVecGen-Plus+ProtVecGen-Ensemble方法,相应的改进分别为+3.38%(BP)和+2.86%(MF)。同样重要的是,在不同长度的蛋白质序列上都获得了稳健的性能。

相似文献

1
MCWS-Transformers: Towards an Efficient Modeling of Protein Sequences via Multi Context-Window Based Scaled Self-Attention.MCWS-Transformer:通过基于多上下文窗口的缩放自注意力实现蛋白质序列的高效建模
IEEE/ACM Trans Comput Biol Bioinform. 2023 Mar-Apr;20(2):1188-1199. doi: 10.1109/TCBB.2022.3173789. Epub 2023 Apr 3.
2
An Ensemble Tf-Idf Based Approach to Protein Function Prediction via Sequence Segmentation.一种基于集成词频-逆文档频率的通过序列分割进行蛋白质功能预测的方法。
IEEE/ACM Trans Comput Biol Bioinform. 2022 Sep-Oct;19(5):2685-2696. doi: 10.1109/TCBB.2021.3093060. Epub 2022 Oct 10.
3
A Sub-Sequence Based Approach to Protein Function Prediction via Multi-Attention Based Multi-Aspect Network.基于子序列的多注意力多方面网络的蛋白质功能预测方法。
IEEE/ACM Trans Comput Biol Bioinform. 2023 Jan-Feb;20(1):94-105. doi: 10.1109/TCBB.2021.3130923. Epub 2023 Feb 3.
4
HN-PPISP: a hybrid network based on MLP-Mixer for protein-protein interaction site prediction.HN-PPISP:一种基于MLP-Mixer的用于蛋白质-蛋白质相互作用位点预测的混合网络。
Brief Bioinform. 2023 Jan 19;24(1). doi: 10.1093/bib/bbac480.
5
LMCrot: an enhanced protein crotonylation site predictor by leveraging an interpretable window-level embedding from a transformer-based protein language model.LMCrot:一种基于转换器的蛋白质语言模型的可解释窗口级嵌入的增强型蛋白质巴豆酰化位点预测器。
Bioinformatics. 2024 May 2;40(5). doi: 10.1093/bioinformatics/btae290.
6
Predicting gene expression levels from DNA sequences and post-transcriptional information with transformers.利用转换器从 DNA 序列和转录后信息预测基因表达水平。
Comput Methods Programs Biomed. 2022 Oct;225:107035. doi: 10.1016/j.cmpb.2022.107035. Epub 2022 Aug 7.
7
Vision Transformer-based recognition of diabetic retinopathy grade.基于 Vision Transformer 的糖尿病视网膜病变分级识别。
Med Phys. 2021 Dec;48(12):7850-7863. doi: 10.1002/mp.15312. Epub 2021 Nov 16.
8
ABT-MPNN: an atom-bond transformer-based message-passing neural network for molecular property prediction.ABT-MPNN:一种基于原子键变压器的消息传递神经网络,用于分子性质预测。
J Cheminform. 2023 Feb 26;15(1):29. doi: 10.1186/s13321-023-00698-9.
9
MESTrans: Multi-scale embedding spatial transformer for medical image segmentation.MESTrans:用于医学图像分割的多尺度嵌入空间变换器
Comput Methods Programs Biomed. 2023 May;233:107493. doi: 10.1016/j.cmpb.2023.107493. Epub 2023 Mar 17.
10
Multi-Modal Adaptive Fusion Transformer Network for the Estimation of Depression Level.多模态自适应融合 Transformer 网络用于抑郁水平估计。
Sensors (Basel). 2021 Jul 12;21(14):4764. doi: 10.3390/s21144764.