Department of Computer Science, University of Virginia, Charlottesville, VA 22903, USA.
Emerging Diagnostic and Investigative Technologies, Department of Pathology and Laboratory Medicine, Dartmouth Health, Lebanon, NH 03766, USA.
Brief Bioinform. 2024 Sep 23;25(6). doi: 10.1093/bib/bbae476.
The application of deep learning to spatial transcriptomics (ST) can reveal relationships between gene expression and tissue architecture. Prior work has demonstrated that inferring gene expression from tissue histomorphology can discern these spatial molecular markers to enable population scale studies, reducing the fiscal barriers associated with large-scale spatial profiling. However, while most improvements in algorithmic performance have focused on improving model architectures, little is known about how the quality of tissue preparation and imaging can affect deep learning model training for spatial inference from morphology and its potential for widespread clinical adoption. Prior studies for ST inference from histology typically utilize manually stained frozen sections with imaging on non-clinical grade scanners. Training such models on ST cohorts is also costly. We hypothesize that adopting tissue processing and imaging practices that mirror standards for clinical implementation (permanent sections, automated tissue staining, and clinical grade scanning) can significantly improve model performance. An enhanced specimen processing and imaging protocol was developed for deep learning-based ST inference from morphology. This protocol featured the Visium CytAssist assay to permit automated hematoxylin and eosin staining (e.g. Leica Bond), 40×-resolution imaging, and joining of multiple patients' tissue sections per capture area prior to ST profiling. Using a cohort of 13 pathologic T Stage-III stage colorectal cancer patients, we compared the performance of models trained on slide prepared using enhanced versus traditional (i.e. manual staining and low-resolution imaging) protocols. Leveraging Inceptionv3 neural networks, we predicted gene expression across serial, histologically-matched tissue sections using whole slide images (WSI) from both protocols. The data Shapley was used to quantify and compare marginal performance gains on a patient-by-patient basis attributed to using the enhanced protocol versus the actual costs of spatial profiling. Findings indicate that training and validating on WSI acquired through the enhanced protocol as opposed to the traditional method resulted in improved performance at lower fiscal cost. In the realm of ST, the enhancement of deep learning architectures frequently captures the spotlight; however, the significance of specimen processing and imaging is often understated. This research, informed through a game-theoretic lens, underscores the substantial impact that specimen preparation/imaging can have on spatial transcriptomic inference from morphology. It is essential to integrate such optimized processing protocols to facilitate the identification of prognostic markers at a larger scale.
深度学习在空间转录组学(ST)中的应用可以揭示基因表达与组织架构之间的关系。先前的研究表明,从组织形态学推断基因表达可以识别这些空间分子标记,从而实现大规模的群体研究,降低与大规模空间分析相关的财政障碍。然而,尽管算法性能的大多数改进都集中在改进模型架构上,但对于组织准备和成像的质量如何影响基于形态学的空间推断的深度学习模型训练及其在广泛的临床应用中的潜力,知之甚少。先前用于从组织学推断 ST 的研究通常利用手动染色的冷冻切片,并在非临床级扫描仪上进行成像。在 ST 队列上训练此类模型的成本也很高。我们假设采用类似于临床实施标准的组织处理和成像实践(永久切片、自动组织染色和临床级扫描)可以显著提高模型性能。我们开发了一种基于深度学习的形态学 ST 推断的增强型标本处理和成像方案。该方案采用 Visium CytAssist 检测试剂盒,以允许自动苏木精和伊红染色(例如 Leica Bond),40×分辨率成像,以及在 ST 分析之前,将多个患者的组织切片连接到每个捕获区域。使用 13 例病理 T 期-III 期结直肠癌患者的队列,我们比较了使用增强型(即手动染色和低分辨率成像)与传统型(即手动染色和低分辨率成像)方案制备的幻灯片上训练的模型的性能。利用 Inceptionv3 神经网络,我们使用两种方案的全幻灯片图像(WSI)预测了连续的、组织学匹配的组织切片中的基因表达。使用数据 Shapley 来量化和比较基于患者的边际性能增益,这些增益归因于使用增强型方案与空间分析的实际成本。结果表明,与传统方法相比,在增强型方案中获取的 WSI 上进行训练和验证可以以较低的财务成本提高性能。在 ST 领域,深度学习架构的增强经常成为焦点;然而,标本处理和成像的重要性往往被低估。这项研究通过博弈论的视角强调了标本制备/成像对形态学空间转录组推断的重大影响。整合这种优化处理方案对于在更大规模上识别预后标志物至关重要。