• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

T细胞受体表征的对比学习

Contrastive learning of T cell receptor representations.

作者信息

Nagano Yuta, Pyo Andrew G T, Milighetti Martina, Henderson James, Shawe-Taylor John, Chain Benny, Tiffeau-Mayer Andreas

机构信息

Division of Infection and Immunity, University College London, London WC1E 6BT, UK; Division of Medicine, University College London, London WC1E 6BT, UK.

Center for the Physics of Biological Function, Princeton University, Princeton, NJ 08544, USA.

出版信息

Cell Syst. 2025 Jan 15;16(1):101165. doi: 10.1016/j.cels.2024.12.006. Epub 2025 Jan 7.

DOI:10.1016/j.cels.2024.12.006
PMID:39778580
Abstract

Computational prediction of the interaction of T cell receptors (TCRs) and their ligands is a grand challenge in immunology. Despite advances in high-throughput assays, specificity-labeled TCR data remain sparse. In other domains, the pre-training of language models on unlabeled data has been successfully used to address data bottlenecks. However, it is unclear how to best pre-train protein language models for TCR specificity prediction. Here, we introduce a TCR language model called SCEPTR (simple contrastive embedding of the primary sequence of T cell receptors), which is capable of data-efficient transfer learning. Through our model, we introduce a pre-training strategy combining autocontrastive learning and masked-language modeling, which enables SCEPTR to achieve its state-of-the-art performance. In contrast, existing protein language models and a variant of SCEPTR pre-trained without autocontrastive learning are outperformed by sequence alignment-based methods. We anticipate that contrastive learning will be a useful paradigm to decode the rules of TCR specificity. A record of this paper's transparent peer review process is included in the supplemental information.

摘要

T细胞受体(TCR)与其配体相互作用的计算预测是免疫学中的一项重大挑战。尽管高通量检测技术取得了进展,但特异性标记的TCR数据仍然稀少。在其他领域,在未标记数据上对语言模型进行预训练已成功用于解决数据瓶颈问题。然而,目前尚不清楚如何最好地对蛋白质语言模型进行预训练以用于TCR特异性预测。在此,我们引入了一种名为SCEPTR(T细胞受体一级序列的简单对比嵌入)的TCR语言模型,它能够进行数据高效的迁移学习。通过我们的模型,我们引入了一种将自对比学习和掩码语言建模相结合的预训练策略,这使得SCEPTR能够实现其目前的最优性能。相比之下,现有的蛋白质语言模型以及未经自对比学习预训练的SCEPTR变体在基于序列比对的方法面前表现较差。我们预计对比学习将成为解码TCR特异性规则的一种有用范式。本文透明的同行评审过程记录包含在补充信息中。

相似文献

1
Contrastive learning of T cell receptor representations.T细胞受体表征的对比学习
Cell Syst. 2025 Jan 15;16(1):101165. doi: 10.1016/j.cels.2024.12.006. Epub 2025 Jan 7.
2
TCR clustering by contrastive learning on antigen specificity.通过对抗原特异性的对比学习进行 TCR 聚类。
Brief Bioinform. 2024 Jul 25;25(5). doi: 10.1093/bib/bbae375.
3
DeepLION2: deep multi-instance contrastive learning framework enhancing the prediction of cancer-associated T cell receptors by attention strategy on motifs.DeepLION2:基于模体注意力策略的深度多实例对比学习框架,用于增强对癌症相关 T 细胞受体的预测。
Front Immunol. 2024 Mar 7;15:1345586. doi: 10.3389/fimmu.2024.1345586. eCollection 2024.
4
Designing meaningful continuous representations of T cell receptor sequences with deep generative models.利用深度生成模型设计有意义的 T 细胞受体序列连续表示。
Nat Commun. 2024 May 20;15(1):4271. doi: 10.1038/s41467-024-48198-0.
5
A comprehensive benchmarking for evaluating TCR embeddings in modeling TCR-epitope interactions.一种用于评估TCR嵌入在模拟TCR-表位相互作用中的综合基准测试。
Brief Bioinform. 2024 Nov 22;26(1). doi: 10.1093/bib/bbaf030.
6
TCR-H: explainable machine learning prediction of T-cell receptor epitope binding on unseen datasets.TCR-H:在未见数据集上解释性机器学习预测 T 细胞受体表位结合
Front Immunol. 2024 Aug 16;15:1426173. doi: 10.3389/fimmu.2024.1426173. eCollection 2024.
7
DECODE: a computational pipeline to discover T cell receptor binding rules.DECODE:一种用于发现 T 细胞受体结合规则的计算管道。
Bioinformatics. 2022 Jun 24;38(Suppl 1):i246-i254. doi: 10.1093/bioinformatics/btac257.
8
NetTCR-2.1: Lessons and guidance on how to develop models for TCR specificity predictions.NetTCR-2.1:关于如何开发 TCR 特异性预测模型的经验教训和指导。
Front Immunol. 2022 Dec 6;13:1055151. doi: 10.3389/fimmu.2022.1055151. eCollection 2022.
9
Interpretable Prediction of SARS-CoV-2 Epitope-Specific TCR Recognition Using a Pre-Trained Protein Language Model.利用预先训练的蛋白质语言模型可对 SARS-CoV-2 表位特异性 TCR 识别进行可解释预测。
IEEE/ACM Trans Comput Biol Bioinform. 2024 May-Jun;21(3):428-438. doi: 10.1109/TCBB.2024.3368046. Epub 2024 Jun 5.
10
Self-supervised learning of T cell receptor sequences exposes core properties for T cell membership.基于 TCR 序列的自监督学习揭示了 T 细胞身份的核心特征。
Sci Adv. 2024 Apr 26;10(17):eadk4670. doi: 10.1126/sciadv.adk4670.

引用本文的文献

1
Computation strategies and clinical applications in neoantigen discovery towards precision cancer immunotherapy.精准癌症免疫治疗新抗原发现中的计算策略与临床应用
Biomark Res. 2025 Jul 9;13(1):96. doi: 10.1186/s40364-025-00808-9.