• 文献检索
  • 文档翻译
  • 深度研究
  • 学术资讯
  • Suppr Zotero 插件Zotero 插件
  • 邀请有礼
  • 套餐&价格
  • 历史记录
应用&插件
Suppr Zotero 插件Zotero 插件浏览器插件Mac 客户端Windows 客户端微信小程序
定价
高级版会员购买积分包购买API积分包
服务
文献检索文档翻译深度研究API 文档MCP 服务
关于我们
关于 Suppr公司介绍联系我们用户协议隐私条款
关注我们

Suppr 超能文献

核心技术专利:CN118964589B侵权必究
粤ICP备2023148730 号-1Suppr @ 2026

文献检索

告别复杂PubMed语法,用中文像聊天一样搜索,搜遍4000万医学文献。AI智能推荐,让科研检索更轻松。

立即免费搜索

文件翻译

保留排版,准确专业,支持PDF/Word/PPT等文件格式,支持 12+语言互译。

免费翻译文档

深度研究

AI帮你快速写综述,25分钟生成高质量综述,智能提取关键信息,辅助科研写作。

立即免费体验

相似文献

1
Self-supervised Contrastive Video-Speech Representation Learning for Ultrasound.用于超声的自监督对比视频-语音表征学习
Med Image Comput Comput Assist Interv. 2020 Oct;12263:534-543. doi: 10.1007/978-3-030-59716-0_51.
2
Dual Representation Learning From Fetal Ultrasound Video And Sonographer Audio.基于胎儿超声视频和超声检查医师音频的双表征学习
Proc IEEE Int Symp Biomed Imaging. 2024 May 27;2024:1-4. doi: 10.1109/ISBI56570.2024.10635693.
3
Self-Supervised Representation Learning for Ultrasound Video.超声视频的自监督表征学习
Proc IEEE Int Symp Biomed Imaging. 2020 Apr 3;2020:1847-1850. doi: 10.1109/ISBI45749.2020.9098666.
4
Audio-visual modelling in a clinical setting.临床环境中的视听建模。
Sci Rep. 2024 Jul 6;14(1):15569. doi: 10.1038/s41598-024-66160-4.
5
Anatomy-Aware Contrastive Representation Learning for Fetal Ultrasound.用于胎儿超声的解剖学感知对比表示学习
Comput Vis ECCV. 2022 Oct;2022:422-436. doi: 10.1007/978-3-031-25066-8_23.
6
Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation.基于伪标签自训练的局部对比损失的半监督医学图像分割。
Med Image Anal. 2023 Jul;87:102792. doi: 10.1016/j.media.2023.102792. Epub 2023 Mar 11.
7
Margin-aware optimized contrastive learning for enhanced self-supervised histopathological image classification.用于增强自监督组织病理学图像分类的边缘感知优化对比学习
Health Inf Sci Syst. 2024 Nov 29;13(1):2. doi: 10.1007/s13755-024-00316-4. eCollection 2025 Dec.
8
Unsupervised Modality-Transferable Video Highlight Detection With Representation Activation Sequence Learning.基于表征激活序列学习的无监督模态可转移视频高光检测
IEEE Trans Image Process. 2024;33:1911-1922. doi: 10.1109/TIP.2024.3372469. Epub 2024 Mar 12.
9
Cross-view motion consistent self-supervised video inter-intra contrastive for action representation understanding.跨视图运动一致的自我监督视频内-外对比动作表示理解。
Neural Netw. 2024 Nov;179:106578. doi: 10.1016/j.neunet.2024.106578. Epub 2024 Jul 26.
10
Boundary-aware information maximization for self-supervised medical image segmentation.用于自监督医学图像分割的边界感知信息最大化
Med Image Anal. 2024 May;94:103150. doi: 10.1016/j.media.2024.103150. Epub 2024 Mar 28.

引用本文的文献

1
Dual Representation Learning From Fetal Ultrasound Video And Sonographer Audio.基于胎儿超声视频和超声检查医师音频的双表征学习
Proc IEEE Int Symp Biomed Imaging. 2024 May 27;2024:1-4. doi: 10.1109/ISBI56570.2024.10635693.
2
A Systematic Review and Implementation Guidelines of Multimodal Foundation Models in Medical Imaging.医学影像中多模态基础模型的系统评价与实施指南
Res Sq. 2025 Apr 28:rs.3.rs-5537908. doi: 10.21203/rs.3.rs-5537908/v1.
3
Audio-visual modelling in a clinical setting.临床环境中的视听建模。
Sci Rep. 2024 Jul 6;14(1):15569. doi: 10.1038/s41598-024-66160-4.
4
Siamese deep learning video flow cytometry for automatic and label-free clinical cervical cancer cell analysis.暹罗深度学习视频流式细胞术用于自动且无标记的临床宫颈癌细胞分析。
Biomed Opt Express. 2024 Mar 4;15(4):2063-2077. doi: 10.1364/BOE.510022. eCollection 2024 Apr 1.
5
Anatomy-Aware Contrastive Representation Learning for Fetal Ultrasound.用于胎儿超声的解剖学感知对比表示学习
Comput Vis ECCV. 2022 Oct;2022:422-436. doi: 10.1007/978-3-031-25066-8_23.
6
Deep clustering for abdominal organ classification in ultrasound imaging.用于超声成像中腹部器官分类的深度聚类
J Med Imaging (Bellingham). 2023 May;10(3):034502. doi: 10.1117/1.JMI.10.3.034502. Epub 2023 May 18.
7
Self-Supervised Contrastive Learning to Predict Alzheimer's Disease Progression with 3D Amyloid-PET.基于3D淀粉样蛋白PET的自监督对比学习预测阿尔茨海默病进展
medRxiv. 2023 Apr 26:2023.04.20.23288886. doi: 10.1101/2023.04.20.23288886.
8
Self-supervised learning for medical image classification: a systematic review and implementation guidelines.用于医学图像分类的自监督学习:系统综述与实施指南
NPJ Digit Med. 2023 Apr 26;6(1):74. doi: 10.1038/s41746-023-00811-0.
9
Audio self-supervised learning: A survey.音频自监督学习:一项综述。
Patterns (N Y). 2022 Dec 9;3(12):100616. doi: 10.1016/j.patter.2022.100616.
10
Contrastive self-supervised learning from 100 million medical images with optional supervision.基于一亿张医学图像的对比自监督学习及可选监督。
J Med Imaging (Bellingham). 2022 Nov;9(6):064503. doi: 10.1117/1.JMI.9.6.064503. Epub 2022 Nov 30.

本文引用的文献

1
Self-Supervised Representation Learning for Ultrasound Video.超声视频的自监督表征学习
Proc IEEE Int Symp Biomed Imaging. 2020 Apr 3;2020:1847-1850. doi: 10.1109/ISBI45749.2020.9098666.
2
Squeeze-and-Excitation Networks.挤压激励网络。
IEEE Trans Pattern Anal Mach Intell. 2020 Aug;42(8):2011-2023. doi: 10.1109/TPAMI.2019.2913372. Epub 2019 Apr 29.
3
What Do Different Evaluation Metrics Tell Us About Saliency Models?不同的评估指标能告诉我们关于显著性模型的哪些信息?
IEEE Trans Pattern Anal Mach Intell. 2019 Mar;41(3):740-757. doi: 10.1109/TPAMI.2018.2815601. Epub 2018 Mar 13.
4
SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound.SonoNet:徒手超声中胎儿标准扫描平面的实时检测与定位
IEEE Trans Med Imaging. 2017 Nov;36(11):2204-2215. doi: 10.1109/TMI.2017.2712367. Epub 2017 Jul 11.

用于超声的自监督对比视频-语音表征学习

Self-supervised Contrastive Video-Speech Representation Learning for Ultrasound.

作者信息

Jiao Jianbo, Cai Yifan, Alsharid Mohammad, Drukker Lior, Papageorghiou Aris T, Noble J Alison

机构信息

Department of Engineering Science, University of Oxford, Oxford, UK.

Nuffield Department of Women's & Reproductive Health, University of Oxford, UK.

出版信息

Med Image Comput Comput Assist Interv. 2020 Oct;12263:534-543. doi: 10.1007/978-3-030-59716-0_51.

DOI:10.1007/978-3-030-59716-0_51
PMID:33103162
原文链接:https://pmc.ncbi.nlm.nih.gov/articles/PMC7116253/
Abstract

In medical imaging, manual annotations can be expensive to acquire and sometimes infeasible to access, making conventional deep learning-based models difficult to scale. As a result, it would be beneficial if useful representations could be derived from raw data without the need for manual annotations. In this paper, we propose to address the problem of self-supervised representation learning with multi-modal ultrasound video-speech raw data. For this case, we assume that there is a high correlation between the ultrasound video and the corresponding narrative speech audio of the sonographer. In order to learn meaningful representations, the model needs to identify such correlation and at the same time understand the underlying anatomical features. We designed a framework to model the correspondence between video and audio without any kind of human annotations. Within this framework, we introduce cross-modal contrastive learning and an affinity-aware self-paced learning scheme to enhance correlation modelling. Experimental evaluations on multi-modal fetal ultrasound video and audio show that the proposed approach is able to learn strong representations and transfers well to downstream tasks of standard plane detection and eye-gaze prediction.

摘要

在医学成像中,手动标注获取成本高昂,有时甚至无法获取,这使得传统的基于深度学习的模型难以扩展。因此,如果能够从原始数据中导出有用的表示而无需手动标注,那将是有益的。在本文中,我们提议解决利用多模态超声视频-语音原始数据进行自监督表示学习的问题。对于这种情况,我们假设超声视频与超声医师相应的叙述语音音频之间存在高度相关性。为了学习有意义的表示,模型需要识别这种相关性,同时理解潜在的解剖特征。我们设计了一个框架,用于在没有任何人工标注的情况下对视频和音频之间的对应关系进行建模。在此框架内,我们引入了跨模态对比学习和亲和力感知自步学习方案,以增强相关性建模。对多模态胎儿超声视频和音频的实验评估表明,所提出的方法能够学习到强大的表示,并能很好地迁移到标准平面检测和目光注视预测等下游任务中。