Xiao Xiao, Kong Yan, Li Ronghan, Wang Zuoheng, Lu Hui
State Key Laboratory of Microbial Metabolism, Joint International Research Laboratory of Metabolic and Developmental Sciences, Department of Bioinformatics and Biostatistics, School of Life Sciences and Biotechnology, Shanghai Jiao Tong University, Shanghai, China; SJTU-Yale Joint Center for Biostatistics and Data Science, National Center for Translational Medicine, MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, Shanghai, China; Department of Biostatistics, Yale School of Public Health, Yale University, New Haven, CT, United States.
State Key Laboratory of Microbial Metabolism, Joint International Research Laboratory of Metabolic and Developmental Sciences, Department of Bioinformatics and Biostatistics, School of Life Sciences and Biotechnology, Shanghai Jiao Tong University, Shanghai, China; SJTU-Yale Joint Center for Biostatistics and Data Science, National Center for Translational Medicine, MoE Key Lab of Artificial Intelligence, AI Institute, Shanghai Jiao Tong University, Shanghai, China.
Med Image Anal. 2024 Jan;91:103040. doi: 10.1016/j.media.2023.103040. Epub 2023 Nov 20.
Inferring gene expressions from histopathological images has long been a fascinating yet challenging task, primarily due to the substantial disparities between the two modality. Existing strategies using local or global features of histological images are suffering model complexity, GPU consumption, low interpretability, insufficient encoding of local features, and over-smooth prediction of gene expressions among neighboring sites. In this paper, we develop TCGN (Transformer with Convolution and Graph-Node co-embedding method) for gene expression estimation from H&E-stained pathological slide images. TCGN comprises a combination of convolutional layers, transformer encoders, and graph neural networks, and is the first to integrate these blocks in a general and interpretable computer vision backbone. Notably, TCGN uniquely operates with just a single spot image as input for histopathological image analysis, simplifying the process while maintaining interpretability. We validate TCGN on three publicly available spatial transcriptomic datasets. TCGN consistently exhibited the best performance (with median PCC 0.232). TCGN offers superior accuracy while keeping parameters to a minimum (just 86.241 million), and it consumes minimal memory, allowing it to run smoothly even on personal computers. Moreover, TCGN can be extended to handle bulk RNA-seq data while providing the interpretability. Enhancing the accuracy of omics information prediction from pathological images not only establishes a connection between genotype and phenotype, enabling the prediction of costly-to-measure biomarkers from affordable histopathological images, but also lays the groundwork for future multi-modal data modeling. Our results confirm that TCGN is a powerful tool for inferring gene expressions from histopathological images in precision health applications.
从组织病理学图像推断基因表达长期以来一直是一项引人入胜但具有挑战性的任务,主要是因为这两种模态之间存在巨大差异。现有的使用组织学图像局部或全局特征的策略存在模型复杂度高、GPU消耗大、可解释性低、局部特征编码不足以及相邻位点基因表达预测过度平滑等问题。在本文中,我们开发了TCGN(卷积与图节点共嵌入的Transformer方法)用于从苏木精-伊红(H&E)染色的病理切片图像估计基因表达。TCGN由卷积层、Transformer编码器和图神经网络组成,并且是首个将这些模块集成到一个通用且可解释的计算机视觉主干中的方法。值得注意的是,TCGN仅以单个斑点图像作为组织病理学图像分析的输入进行独特操作,在保持可解释性的同时简化了流程。我们在三个公开可用的空间转录组数据集上验证了TCGN。TCGN始终表现出最佳性能(中位数皮尔逊相关系数为0.232)。TCGN在将参数保持在最低水平(仅8624.1万个)的同时提供了卓越的准确性,并且消耗的内存最少,甚至可以在个人计算机上平稳运行。此外,TCGN可以扩展以处理批量RNA测序数据,同时提供可解释性。提高病理图像中组学信息预测的准确性不仅建立了基因型和表型之间的联系,使得能够从经济实惠的组织病理学图像预测成本高昂的生物标志物,而且还为未来的多模态数据建模奠定了基础。我们的结果证实,TCGN是在精准健康应用中从组织病理学图像推断基因表达的强大工具。