Caughlin Kayla, Duran-Sierra Elvis, Cheng Shuna, Cuenca Rodrigo, Ahmed Beena, Ji Jim, Martinez Mathias, Al-Khalil Moustafa, Al-Enazi Hussain, Jo Javier A, Busso Carlos
Department of Electrical and Computer Engineering, The University of Texas at Dallas, Richardson, TX 75080, USA.
Department of Biomedical Engineering, Texas A&M University, College Station, TX 77840, USA.
Cancers (Basel). 2024 Dec 9;16(23):4120. doi: 10.3390/cancers16234120.
Multispectral autofluorescence lifetime imaging systems have recently been developed to quickly and non-invasively assess tissue properties for applications in oral cancer diagnosis. As a non-traditional imaging modality, the autofluorescence signal collected from the system cannot be directly visually assessed by a clinician and a model is needed to generate a diagnosis for each image. However, training a deep learning model from scratch on small multispectral autofluorescence datasets can fail due to inter-patient variability, poor initialization, and overfitting. We propose a contrastive-based pre-training approach that teaches the network to perform patient normalization without requiring a direct comparison to a reference sample. We then use the contrastive pre-trained encoder as a favorable initialization for classification. To train the classifiers, we efficiently use available data and reduce overfitting through a multitask framework with margin delineation and cancer diagnosis tasks. We evaluate the model over 67 patients using 10-fold cross-validation and evaluate significance using paired, one-tailed -tests. The proposed approach achieves a sensitivity of 82.08% and specificity of 75.92% on the cancer diagnosis task with a sensitivity of 91.83% and specificity of 79.31% for margin delineation as an auxiliary task. In comparison to existing approaches, our method significantly outperforms a (SVM) implemented with either (SFS) ( = 0.0261) or L1 loss ( = 0.0452) when considering the average of sensitivity and specificity. Specifically, the proposed approach increases performance by 2.75% compared to the L1 model and 4.87% compared to the SFS model. In addition, there is a significant increase in specificity of 8.34% compared to the baseline autoencoder model ( = 0.0070). Our method effectively trains deep learning models for small data applications when existing, large pre-trained models are not suitable for fine-tuning. While we designed the network for a specific imaging modality, we report the development process so that the insights gained can be applied to address similar challenges in other non-traditional imaging modalities. A key contribution of this paper is a neural network framework for multi-spectral fluorescence lifetime-based tissue discrimination that performs patient normalization without requiring a reference (healthy) sample from each patient at test time.
多光谱自体荧光寿命成像系统最近已被开发出来,用于在口腔癌诊断中快速、无创地评估组织特性。作为一种非传统的成像方式,从该系统收集的自体荧光信号无法由临床医生直接进行视觉评估,因此需要一个模型来对每张图像进行诊断。然而,在小型多光谱自体荧光数据集上从头开始训练深度学习模型可能会失败,原因包括患者间的变异性、初始化不佳和过拟合。我们提出了一种基于对比的预训练方法,该方法能让网络在无需与参考样本直接比较的情况下对患者进行归一化处理。然后,我们将经过对比预训练的编码器用作分类的良好初始化。为了训练分类器,我们有效利用可用数据,并通过带有边界划分和癌症诊断任务的多任务框架减少过拟合。我们使用10折交叉验证对67名患者的模型进行评估,并使用配对单尾检验评估显著性。所提出的方法在癌症诊断任务上的灵敏度为82.08%,特异性为75.92%,作为辅助任务的边界划分的灵敏度为91.83%,特异性为79.31%。与现有方法相比,在考虑灵敏度和特异性的平均值时,我们的方法显著优于使用顺序前向选择(SFS)(p = 0.0261)或L1损失(p = 0.0452)实现的支持向量机(SVM)。具体而言,与L1模型相比,所提出的方法性能提高了2.75%,与SFS模型相比提高了4.87%。此外,与基线自动编码器模型相比,特异性显著提高了8.34%(p = 0.0070)。当现有的大型预训练模型不适合微调时,我们的方法能有效地为小数据应用训练深度学习模型。虽然我们为特定的成像方式设计了网络,但我们报告了开发过程,以便所获得的见解可用于应对其他非传统成像方式中的类似挑战。本文的一个关键贡献是一个基于多光谱荧光寿命的组织鉴别神经网络框架,该框架在测试时无需每个患者的参考(健康)样本即可对患者进行归一化处理。