Suppr超能文献

通过监督学习或自监督学习对胸部X光片分类进行特定领域预训练的再思考:与冷启动主动学习中的ImageNet对应模型的比较研究

Rethinking Domain-Specific Pretraining by Supervised or Self-Supervised Learning for Chest Radiograph Classification: A Comparative Study Against ImageNet Counterparts in Cold-Start Active Learning.

作者信息

Yuan Han, Zhu Mingcheng, Yang Rui, Liu Han, Li Irene, Hong Chuan

机构信息

Duke-NUS Medical School, Centre for Quantitative Medicine Singapore Singapore.

Department of Engineering Science University of Oxford Oxford UK.

出版信息

Health Care Sci. 2025 Apr 6;4(2):110-143. doi: 10.1002/hcs2.70009. eCollection 2025 Apr.

Abstract

OBJECTIVE

Deep learning (DL) has become the prevailing method in chest radiograph analysis, yet its performance heavily depends on large quantities of annotated images. To mitigate the cost, cold-start active learning (AL), comprising an initialization followed by subsequent learning, selects a small subset of informative data points for labeling. Recent advancements in pretrained models by supervised or self-supervised learning tailored to chest radiograph have shown broad applicability to diverse downstream tasks. However, their potential in cold-start AL remains unexplored.

METHODS

To validate the efficacy of domain-specific pretraining, we compared two foundation models: supervised TXRV and self-supervised REMEDIS with their general domain counterparts pretrained on ImageNet. Model performance was evaluated at both initialization and subsequent learning stages on two diagnostic tasks: psychiatric pneumonia and COVID-19. For initialization, we assessed their integration with three strategies: diversity, uncertainty, and hybrid sampling. For subsequent learning, we focused on uncertainty sampling powered by different pretrained models. We also conducted statistical tests to compare the foundation models with ImageNet counterparts, investigate the relationship between initialization and subsequent learning, examine the performance of one-shot initialization against the full AL process, and investigate the influence of class balance in initialization samples on initialization and subsequent learning.

RESULTS

First, domain-specific foundation models failed to outperform ImageNet counterparts in six out of eight experiments on informative sample selection. Both domain-specific and general pretrained models were unable to generate representations that could substitute for the original images as model inputs in seven of the eight scenarios. However, pretrained model-based initialization surpassed random sampling, the default approach in cold-start AL. Second, initialization performance was positively correlated with subsequent learning performance, highlighting the importance of initialization strategies. Third, one-shot initialization performed comparably to the full AL process, demonstrating the potential of reducing experts' repeated waiting during AL iterations. Last, a U-shaped correlation was observed between the class balance of initialization samples and model performance, suggesting that the class balance is more strongly associated with performance at middle budget levels than at low or high budgets.

CONCLUSIONS

In this study, we highlighted the limitations of medical pretraining compared to general pretraining in the context of cold-start AL. We also identified promising outcomes related to cold-start AL, including initialization based on pretrained models, the positive influence of initialization on subsequent learning, the potential for one-shot initialization, and the influence of class balance on middle-budget AL. Researchers are encouraged to improve medical pretraining for versatile DL foundations and explore novel AL methods.

摘要

目的

深度学习(DL)已成为胸部X光片分析的主流方法,但其性能在很大程度上依赖于大量带注释的图像。为了降低成本,冷启动主动学习(AL)包括初始化和后续学习,它选择一小部分信息丰富的数据点进行标注。针对胸部X光片的有监督或自监督学习的预训练模型的最新进展已显示出对各种下游任务具有广泛适用性。然而,它们在冷启动主动学习中的潜力仍未得到探索。

方法

为了验证特定领域预训练的有效性,我们比较了两个基础模型:有监督的TXRV和自监督的REMEDIS,以及它们在ImageNet上预训练的通用领域对应模型。在精神性肺炎和新冠肺炎这两项诊断任务的初始化和后续学习阶段对模型性能进行了评估。对于初始化,我们评估了它们与三种策略的整合情况:多样性、不确定性和混合采样。对于后续学习,我们重点关注由不同预训练模型驱动的不确定性采样。我们还进行了统计测试,以比较基础模型与ImageNet对应模型,研究初始化与后续学习之间的关系,检验一次性初始化相对于完整主动学习过程的性能,并研究初始化样本中的类别平衡对初始化和后续学习的影响。

结果

首先,在八项关于信息样本选择的实验中,特定领域的基础模型在六项实验中未能超过ImageNet对应模型。在八种情况中的七种情况下,特定领域和通用预训练模型都无法生成可以替代原始图像作为模型输入的表示。然而,基于预训练模型的初始化超过了冷启动主动学习中的默认方法——随机采样。其次,初始化性能与后续学习性能呈正相关,突出了初始化策略的重要性。第三,一次性初始化的表现与完整主动学习过程相当,这表明在主动学习迭代过程中减少专家重复等待时间具有潜力。最后,在初始化样本的类别平衡与模型性能之间观察到一种U形相关性,这表明类别平衡在中等预算水平下比在低预算或高预算水平下与性能的关联更强。

结论

在本研究中,我们强调了在冷启动主动学习背景下,与通用预训练相比,医学预训练的局限性。我们还确定了与冷启动主动学习相关的有前景的成果,包括基于预训练模型的初始化、初始化对后续学习的积极影响、一次性初始化的潜力以及类别平衡对中等预算主动学习的影响。鼓励研究人员改进用于通用深度学习基础的医学预训练,并探索新的主动学习方法。

https://cdn.ncbi.nlm.nih.gov/pmc/blobs/d371/11997468/f488b6b93498/HCS2-4-110-g005.jpg

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验