• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

用于转录因子结合位点预测的双分支网络对局部特征和全局表示的协同作用

Cooperation of local features and global representations by a dual-branch network for transcription factor binding sites prediction.

作者信息

Yu Yutong, Ding Pengju, Gao Hongli, Liu Guozhu, Zhang Fa, Yu Bin

机构信息

College of Information Science and Technology, Qingdao University of Science and Technology, China.

College of Mathematics and Physics, Qingdao University of Science and Technology, China.

出版信息

Brief Bioinform. 2023 Mar 19;24(2). doi: 10.1093/bib/bbad036.

DOI:10.1093/bib/bbad036
PMID:36748992
Abstract

Interactions between DNA and transcription factors (TFs) play an essential role in understanding transcriptional regulation mechanisms and gene expression. Due to the large accumulation of training data and low expense, deep learning methods have shown huge potential in determining the specificity of TFs-DNA interactions. Convolutional network-based and self-attention network-based methods have been proposed for transcription factor binding sites (TFBSs) prediction. Convolutional operations are efficient to extract local features but easy to ignore global information, while self-attention mechanisms are expert in capturing long-distance dependencies but difficult to pay attention to local feature details. To discover comprehensive features for a given sequence as far as possible, we propose a Dual-branch model combining Self-Attention and Convolution, dubbed as DSAC, which fuses local features and global representations in an interactive way. In terms of features, convolution and self-attention contribute to feature extraction collaboratively, enhancing the representation learning. In terms of structure, a lightweight but efficient architecture of network is designed for the prediction, in particular, the dual-branch structure makes the convolution and the self-attention mechanism can be fully utilized to improve the predictive ability of our model. The experiment results on 165 ChIP-seq datasets show that DSAC obviously outperforms other five deep learning based methods and demonstrate that our model can effectively predict TFBSs based on sequence feature alone. The source code of DSAC is available at https://github.com/YuBinLab-QUST/DSAC/.

摘要

DNA与转录因子(TFs)之间的相互作用在理解转录调控机制和基因表达方面起着至关重要的作用。由于训练数据的大量积累和低成本,深度学习方法在确定TFs与DNA相互作用的特异性方面显示出巨大潜力。基于卷积网络和基于自注意力网络的方法已被提出用于转录因子结合位点(TFBSs)预测。卷积操作有效地提取局部特征,但容易忽略全局信息,而自注意力机制擅长捕捉长距离依赖关系,但难以关注局部特征细节。为了尽可能发现给定序列的综合特征,我们提出了一种结合自注意力和卷积的双分支模型,称为DSAC,它以交互方式融合局部特征和全局表示。在特征方面,卷积和自注意力协同促进特征提取,增强表示学习。在结构方面,设计了一种轻量级但高效的网络架构用于预测,特别是双分支结构使得卷积和自注意力机制能够充分利用,提高我们模型的预测能力。在165个ChIP-seq数据集上的实验结果表明,DSAC明显优于其他五种基于深度学习的方法,并证明我们的模型仅基于序列特征就能有效地预测TFBSs。DSAC的源代码可在https://github.com/YuBinLab-QUST/DSAC/获取。

相似文献

1
Cooperation of local features and global representations by a dual-branch network for transcription factor binding sites prediction.用于转录因子结合位点预测的双分支网络对局部特征和全局表示的协同作用
Brief Bioinform. 2023 Mar 19;24(2). doi: 10.1093/bib/bbad036.
2
DeepSTF: predicting transcription factor binding sites by interpretable deep neural networks combining sequence and shape.DeepSTF:通过结合序列和形状的可解释深度神经网络预测转录因子结合位点。
Brief Bioinform. 2023 Jul 20;24(4). doi: 10.1093/bib/bbad231.
3
A novel convolution attention model for predicting transcription factor binding sites by combination of sequence and shape.一种新型卷积注意力模型,通过序列和形状的结合来预测转录因子结合位点。
Brief Bioinform. 2022 Jan 17;23(1). doi: 10.1093/bib/bbab525.
4
BERT-TFBS: a novel BERT-based model for predicting transcription factor binding sites by transfer learning.BERT-TFBS:一种基于迁移学习的用于预测转录因子结合位点的新型基于BERT的模型。
Brief Bioinform. 2024 Mar 27;25(3). doi: 10.1093/bib/bbae195.
5
Prediction of Transcription Factor Binding Sites With an Attention Augmented Convolutional Neural Network.基于注意力增强卷积神经网络的转录因子结合位点预测
IEEE/ACM Trans Comput Biol Bioinform. 2022 Nov-Dec;19(6):3614-3623. doi: 10.1109/TCBB.2021.3126623. Epub 2022 Dec 8.
6
TBCA: Prediction of transcription factor binding sites using a deep neural network with lightweight attention mechanism.TBCA:使用具有轻量级注意力机制的深度神经网络预测转录因子结合位点
IEEE J Biomed Health Inform. 2024 Jan 18;PP. doi: 10.1109/JBHI.2024.3355758.
7
CRMSNet: A deep learning model that uses convolution and residual multi-head self-attention block to predict RBPs for RNA sequence.CRMSNet:一种深度学习模型,使用卷积和残差多头自注意力块来预测 RNA 序列的 RBPs。
Proteins. 2023 Aug;91(8):1032-1041. doi: 10.1002/prot.26489. Epub 2023 Mar 28.
8
HN-PPISP: a hybrid network based on MLP-Mixer for protein-protein interaction site prediction.HN-PPISP:一种基于MLP-Mixer的用于蛋白质-蛋白质相互作用位点预测的混合网络。
Brief Bioinform. 2023 Jan 19;24(1). doi: 10.1093/bib/bbac480.
9
MLSNet: a deep learning model for predicting transcription factor binding sites.MLSNet:一种用于预测转录因子结合位点的深度学习模型。
Brief Bioinform. 2024 Sep 23;25(6). doi: 10.1093/bib/bbae489.
10
Conformer: Local Features Coupling Global Representations for Recognition and Detection.构象:用于识别和检测的局部特征与全局表示相结合。
IEEE Trans Pattern Anal Mach Intell. 2023 Aug;45(8):9454-9468. doi: 10.1109/TPAMI.2023.3243048. Epub 2023 Jun 30.

引用本文的文献

1
HemoFuse: multi-feature fusion based on multi-head cross-attention for identification of hemolytic peptides.HemoFuse:基于多头交叉注意力的多特征融合用于识别溶血肽。
Sci Rep. 2024 Sep 28;14(1):22518. doi: 10.1038/s41598-024-74326-3.
2
BERT-TFBS: a novel BERT-based model for predicting transcription factor binding sites by transfer learning.BERT-TFBS:一种基于迁移学习的用于预测转录因子结合位点的新型基于BERT的模型。
Brief Bioinform. 2024 Mar 27;25(3). doi: 10.1093/bib/bbae195.
3
AACFlow: an end-to-end model based on attention augmented convolutional neural network and flow-attention mechanism for identification of anticancer peptides.
AACFlow:一种基于注意力增强卷积神经网络和流注意力机制的端到端模型,用于识别抗癌肽。
Bioinformatics. 2024 Mar 4;40(3). doi: 10.1093/bioinformatics/btae142.