• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

DeepNeuropePred:一种通过蛋白质语言模型从神经肽前体预测切割位点的强大通用工具。

DeepNeuropePred: A robust and universal tool to predict cleavage sites from neuropeptide precursors by protein language model.

作者信息

Wang Lei, Zeng Zilu, Xue Zhidong, Wang Yan

机构信息

Institute of Medical Artificial Intelligence, Binzhou Medical University, Yantai, Shandong 264003, China.

School of Life Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei 430074, China.

出版信息

Comput Struct Biotechnol J. 2023 Dec 5;23:309-315. doi: 10.1016/j.csbj.2023.12.004. eCollection 2024 Dec.

DOI:10.1016/j.csbj.2023.12.004
PMID:38179071
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC10764246/
Abstract

Neuropeptides play critical roles in many biological processes such as growth, learning, memory, metabolism, and neuronal differentiation. A few approaches have been reported for predicting neuropeptides that are cleaved from precursor protein sequences. However, these models for cleavage site prediction of precursors were developed using a limited number of neuropeptide precursor datasets and simple precursors representation models. In addition, a universal method for predicting neuropeptide cleavage sites that can be applied to all species is still lacking. In this paper, we proposed a novel deep learning method called DeepNeuropePred, using a combination of pre-trained language model and Convolutional Neural Networks for feature extraction and predicting the neuropeptide cleavage sites from precursors. To demonstrate the model's effectiveness and robustness, we evaluated the performance of DeepNeuropePred and four models from the NeuroPred server in the independent dataset and our model achieved the highest AUC score (0.916), which are 6.9%, 7.8%, 8.8%, and 10.9% higher than Mammalian (0.857), insects (0.850), Mollusc (0.842) and Motif (0.826), respectively. For the convenience of researchers, we provide a web server (http://isyslab.info/NeuroPepV2/deepNeuropePred.jsp).

摘要

神经肽在许多生物过程中发挥着关键作用,如生长、学习、记忆、新陈代谢和神经元分化。已经报道了一些用于预测从前体蛋白序列中切割出来的神经肽的方法。然而,这些用于预测前体切割位点的模型是使用有限数量的神经肽前体数据集和简单的前体表示模型开发的。此外,仍然缺乏一种可应用于所有物种的预测神经肽切割位点的通用方法。在本文中,我们提出了一种名为DeepNeuropePred的新型深度学习方法,它结合了预训练语言模型和卷积神经网络进行特征提取,并从前体中预测神经肽切割位点。为了证明该模型的有效性和稳健性,我们在独立数据集中评估了DeepNeuropePred和来自NeuroPred服务器的四个模型的性能,我们的模型获得了最高的AUC分数(0.916),分别比哺乳动物模型(0.857)、昆虫模型(0.850)、软体动物模型(0.842)和Motif模型(0.826)高6.9%、7.8%、8.8%和10.9%。为方便研究人员,我们提供了一个网络服务器(http://isyslab.info/NeuroPepV2/deepNeuropePred.jsp)。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/042c669de59f/gr4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/c77a80992a8d/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/6c28197d9846/gr2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/95a6d237fbf1/gr3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/042c669de59f/gr4.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/c77a80992a8d/gr1.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/6c28197d9846/gr2.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/95a6d237fbf1/gr3.jpg
https://cdn.ncbi.nlm.nih.gov/pmc/blobs/f23e/10764246/042c669de59f/gr4.jpg

相似文献

1
DeepNeuropePred: A robust and universal tool to predict cleavage sites from neuropeptide precursors by protein language model.DeepNeuropePred:一种通过蛋白质语言模型从神经肽前体预测切割位点的强大通用工具。
Comput Struct Biotechnol J. 2023 Dec 5;23:309-315. doi: 10.1016/j.csbj.2023.12.004. eCollection 2024 Dec.
2
NeuroPred-PLM: an interpretable and robust model for neuropeptide prediction by protein language model.NeuroPred-PLM:一种基于蛋白质语言模型的用于神经肽预测的可解释且稳健的模型。
Brief Bioinform. 2023 Mar 19;24(2). doi: 10.1093/bib/bbad077.
3
NeuroCS: A Tool to Predict Cleavage Sites of Neuropeptide Precursors.NeuroCS:一种预测神经肽前体切割位点的工具。
Protein Pept Lett. 2020;27(4):337-345. doi: 10.2174/0929866526666191112150636.
4
NeuroPep 2.0: An Updated Database Dedicated to Neuropeptide and Its Receptor Annotations.NeuroPep 2.0:一个专注于神经肽及其受体注释的更新数据库。
J Mol Biol. 2024 Feb 15;436(4):168416. doi: 10.1016/j.jmb.2023.168416. Epub 2023 Dec 22.
5
NeuroPred: a tool to predict cleavage sites in neuropeptide precursors and provide the masses of the resulting peptides.NeuroPred:一种预测神经肽前体中切割位点并提供所得肽质量的工具。
Nucleic Acids Res. 2006 Jul 1;34(Web Server issue):W267-72. doi: 10.1093/nar/gkl161.
6
Comparative analysis of neuropeptide cleavage sites in human, mouse, rat, and cattle.人类、小鼠、大鼠和牛神经肽切割位点的比较分析。
Mamm Genome. 2008 Feb;19(2):106-20. doi: 10.1007/s00335-007-9090-9. Epub 2008 Jan 23.
7
A python analytical pipeline to identify prohormone precursors and predict prohormone cleavage sites.一种用于鉴定前体激素原和预测前体激素切割位点的 Python 分析流程。
Front Neuroinform. 2008 Dec 16;2:7. doi: 10.3389/neuro.11.007.2008. eCollection 2008.
8
NeuroPred-FRL: an interpretable prediction model for identifying neuropeptide using feature representation learning.神经肽预测模型 FRL:基于特征表示学习的神经肽识别可解释预测模型。
Brief Bioinform. 2021 Nov 5;22(6). doi: 10.1093/bib/bbab167.
9
Mono- and dibasic proteolytic cleavage sites in insect neuroendocrine peptide precursors.昆虫神经内分泌肽前体中的单碱性和双碱性蛋白水解切割位点。
Arch Insect Biochem Physiol. 2000 Feb;43(2):49-63. doi: 10.1002/(SICI)1520-6327(200002)43:2<49::AID-ARCH1>3.0.CO;2-M.
10
Characterization of the prohormone complement in cattle using genomic libraries and cleavage prediction approaches.利用基因组文库和裂解预测方法对牛前激素补体进行表征。
BMC Genomics. 2009 May 16;10:228. doi: 10.1186/1471-2164-10-228.

引用本文的文献

1
Discovery of peptides as key regulators of metabolic and cardiovascular crosstalk.肽作为代谢与心血管相互作用关键调节因子的发现。
Cell Rep. 2025 Jun 24;44(6):115836. doi: 10.1016/j.celrep.2025.115836. Epub 2025 Jun 16.
2
NeuroScale: evolutional scale-based protein language models enable prediction of neuropeptides.NeuroScale:基于进化尺度的蛋白质语言模型可实现神经肽预测。
BMC Biol. 2025 May 28;23(1):142. doi: 10.1186/s12915-025-02243-6.
3
Understanding peptide hormones: from precursor proteins to bioactive molecules.了解肽类激素:从前体蛋白到生物活性分子。

本文引用的文献

1
NeuroPred-PLM: an interpretable and robust model for neuropeptide prediction by protein language model.NeuroPred-PLM:一种基于蛋白质语言模型的用于神经肽预测的可解释且稳健的模型。
Brief Bioinform. 2023 Mar 19;24(2). doi: 10.1093/bib/bbad077.
2
Res-Dom: predicting protein domain boundary from sequence using deep residual network and Bi-LSTM.Res-Dom:使用深度残差网络和双向长短期记忆网络从序列预测蛋白质结构域边界
Bioinform Adv. 2022 Sep 1;2(1):vbac060. doi: 10.1093/bioadv/vbac060. eCollection 2022.
3
DistilProtBert: a distilled protein language model used to distinguish between real proteins and their randomly shuffled counterparts.
Trends Biochem Sci. 2025 Jun;50(6):481-494. doi: 10.1016/j.tibs.2025.03.014. Epub 2025 Apr 14.
4
Computational approaches for identifying neuropeptides: A comprehensive review.识别神经肽的计算方法:综述
Mol Ther Nucleic Acids. 2024 Nov 28;36(1):102409. doi: 10.1016/j.omtn.2024.102409. eCollection 2025 Mar 11.
DistilProtBert:一种经过蒸馏的蛋白质语言模型,用于区分真实蛋白质与其随机打乱的对应物。
Bioinformatics. 2022 Sep 16;38(Suppl_2):ii95-ii98. doi: 10.1093/bioinformatics/btac474.
4
Identify Bitter Peptides by Using Deep Representation Learning Features.利用深度表示学习特征识别苦味肽。
Int J Mol Sci. 2022 Jul 17;23(14):7877. doi: 10.3390/ijms23147877.
5
Improving the topology prediction of α-helical transmembrane proteins with deep transfer learning.利用深度迁移学习改进α-螺旋跨膜蛋白的拓扑结构预测
Comput Struct Biotechnol J. 2022 Apr 20;20:1993-2000. doi: 10.1016/j.csbj.2022.04.024. eCollection 2022.
6
DeepLoc 2.0: multi-label subcellular localization prediction using protein language models.DeepLoc 2.0:使用蛋白质语言模型进行多标签亚细胞定位预测。
Nucleic Acids Res. 2022 Jul 5;50(W1):W228-W234. doi: 10.1093/nar/gkac278.
7
SignalP 6.0 predicts all five types of signal peptides using protein language models.SignalP 6.0 使用蛋白质语言模型预测所有五种类型的信号肽。
Nat Biotechnol. 2022 Jul;40(7):1023-1025. doi: 10.1038/s41587-021-01156-3. Epub 2022 Jan 3.
8
ProtTrans: Toward Understanding the Language of Life Through Self-Supervised Learning.ProtTrans:通过自监督学习理解生命语言。
IEEE Trans Pattern Anal Mach Intell. 2022 Oct;44(10):7112-7127. doi: 10.1109/TPAMI.2021.3095381. Epub 2022 Sep 14.
9
Learning the protein language: Evolution, structure, and function.学习蛋白质语言:进化、结构和功能。
Cell Syst. 2021 Jun 16;12(6):654-669.e3. doi: 10.1016/j.cels.2021.05.017.
10
Biological structure and function emerge from scaling unsupervised learning to 250 million protein sequences.生物结构和功能源于将无监督学习扩展到 2.5 亿个蛋白质序列。
Proc Natl Acad Sci U S A. 2021 Apr 13;118(15). doi: 10.1073/pnas.2016239118.