Suppr超能文献

通过自监督从解剖学中学习局部可定位性、可组合性和可分解性,在基础模型中表示部分-整体层次结构。

Representing Part-Whole Hierarchies in Foundation Models by Learning Localizability, Composability, and Decomposability from Anatomy via Self-Supervision.

作者信息

Taher Mohammad Reza Hosseinzadeh, Gotway Michael B, Liang Jianming

机构信息

Arizona State University.

Mayo Clinic.

出版信息

Proc IEEE Comput Soc Conf Comput Vis Pattern Recognit. 2024 Jun;abs/210504906(2024):11269-11281. doi: 10.1109/cvpr52733.2024.01071. Epub 2024 Sep 16.

Abstract

Humans effortlessly interpret images by parsing them into part-whole hierarchies; deep learning excels in learning multi-level feature spaces, but they often lack explicit coding of part-whole relations, a prominent property of medical imaging. To overcome this limitation, we introduce Adam-v2, a new self-supervised learning framework extending Adam [79] by explicitly incorporating part-whole hierarchies into its learning objectives through three key branches: (1) Localizability, acquiring discriminative representations to distinguish different anatomical patterns; (2) Composability, learning each anatomical structure in a parts-to-whole manner; and (3) Decomposability, comprehending each anatomical structure in a whole-to-parts manner. Experimental results across 10 tasks, compared to 11 baselines in zero-shot, few-shot transfer, and full fine-tuning settings, showcase Adam-v2's superior performance over large-scale medical models and existing SSL methods across diverse downstream tasks. The higher generality and robustness of Adam-v2's representations originate from its explicit construction of hierarchies for distinct anatomical structures from unlabeled medical images. Adam-v2 preserves a semantic balance of anatomical diversity and harmony in its embedding, yielding representations that are both generic and semantically meaningful, yet overlooked in existing SSL methods. All code and pretrained models are available at GitHub.com/JLiangLab/Eden.

摘要

人类通过将图像解析为部分-整体层次结构来轻松地解释图像;深度学习擅长学习多级特征空间,但它们通常缺乏对部分-整体关系的显式编码,而这是医学成像的一个突出特性。为了克服这一局限性,我们引入了Adam-v2,这是一种新的自监督学习框架,它通过三个关键分支将部分-整体层次结构明确纳入其学习目标,从而扩展了Adam [79]:(1)可定位性,获取判别性表示以区分不同的解剖模式;(2)可组合性,以从部分到整体的方式学习每个解剖结构;(3)可分解性,以从整体到部分的方式理解每个解剖结构。在零样本、少样本迁移和全量微调设置下,与11个基线相比,在10个任务上的实验结果表明,Adam-v2在各种下游任务上的性能优于大规模医学模型和现有的自监督学习方法。Adam-v2表示的更高通用性和鲁棒性源于其从未标记医学图像中为不同解剖结构显式构建层次结构。Adam-v2在其嵌入中保持了解剖多样性和协调性的语义平衡,产生了既通用又具有语义意义的表示,但在现有的自监督学习方法中却被忽视了。所有代码和预训练模型可在GitHub.com/JLiangLab/Eden上获取。

相似文献

2
Towards Foundation Models Learned from Anatomy in Medical Imaging via Self-supervision.迈向通过自监督从医学成像解剖学中学习的基础模型。
Domain Adapt Represent Transf (2023). 2024;14293:94-104. doi: 10.1007/978-3-031-45857-6_10. Epub 2023 Oct 14.

本文引用的文献

2
Towards Foundation Models Learned from Anatomy in Medical Imaging via Self-supervision.迈向通过自监督从医学成像解剖学中学习的基础模型。
Domain Adapt Represent Transf (2023). 2024;14293:94-104. doi: 10.1007/978-3-031-45857-6_10. Epub 2023 Oct 14.
9
Discriminative, Restorative, and Adversarial Learning: Stepwise Incremental Pretraining.判别式、恢复式和对抗式学习:逐步增量预训练
Domain Adapt Represent Transf (2022). 2022 Sep;13542:66-76. doi: 10.1007/978-3-031-16852-9_7. Epub 2022 Sep 15.

文献AI研究员

20分钟写一篇综述,助力文献阅读效率提升50倍。

立即体验

用中文搜PubMed

大模型驱动的PubMed中文搜索引擎

马上搜索

文档翻译

学术文献翻译模型,支持多种主流文档格式。

立即体验