Irfan Muhammad, Haq Ijaz Ul, Malik Khalid Mahmood, Muhammad Khan
SMILES LAB, College of Innovation & Technology, University of Michigan-Flint, Flint, MI 48502, USA.
SMILES LAB, College of Innovation & Technology, University of Michigan-Flint, Flint, MI 48502, USA.
Comput Med Imaging Graph. 2025 Jun;122:102507. doi: 10.1016/j.compmedimag.2025.102507. Epub 2025 Feb 9.
Generalizability is one of the biggest challenges hindering the advancement of medical sensing technologies across multiple imaging modalities. This issue is further impaired when the imaging data is limited in scope or of poor quality. To tackle this, we propose a generalized and robust, lightweight one-shot learning method for medical image classification across various imaging modalities, including X-ray, microscopic, and CT scans. Our model introduces a collaborative one-shot training (COST) approach, incorporating both meta-learning and metric-learning. This approach allows for effective training on only one image per class. To ensure generalization with fewer epochs, we employ gradient generalization at dense and fully connected layers, utilizing a lightweight Siamese network with triplet loss and shared parameters. The proposed model was evaluated on 12 medical image datasets from MedMNIST2D, achieving an average accuracy of 91.5 % and area under the curve (AUC) of 0.89, outperforming state-of-the-art models such as ResNet-50 and AutoML by over 10 % on certain datasets. Further, in the OCTMNIST dataset, our model achieved an AUC of 0.91 compared to ResNet-50's 0.77. Ablation studies further validate the superiority of our approach, with the COST method showing significant improvement in convergence speed and accuracy when compared to traditional one-shot learning setups. Additionally, our model's lightweight architecture requires only 0.15 million parameters, making it well-suited for deployment on resource-constrained devices.
通用性是阻碍医学传感技术在多种成像模式中取得进展的最大挑战之一。当成像数据的范围有限或质量较差时,这个问题会进一步恶化。为了解决这个问题,我们提出了一种通用、强大且轻量级的一次性学习方法,用于跨各种成像模式(包括X射线、显微镜和CT扫描)的医学图像分类。我们的模型引入了一种协作式一次性训练(COST)方法,融合了元学习和度量学习。这种方法允许仅对每个类别中的一张图像进行有效训练。为了确保在较少轮次下实现通用性,我们在密集层和全连接层采用梯度泛化,利用具有三元组损失和共享参数的轻量级孪生网络。所提出的模型在来自MedMNIST2D的12个医学图像数据集上进行了评估,平均准确率达到91.5%,曲线下面积(AUC)为0.89,在某些数据集上比诸如ResNet-50和AutoML等现有模型的性能高出10%以上。此外,在OCTMNIST数据集中,我们的模型AUC达到0.91,而ResNet-50为0.77。消融研究进一步验证了我们方法的优越性,与传统的一次性学习设置相比,COST方法在收敛速度和准确率方面有显著提高。此外,我们模型的轻量级架构仅需要15万个参数,非常适合在资源受限的设备上进行部署。