Shin Jae-Man, Park Seongyong, Shin Keewon, Seo Woo-Young, Kim Hyun-Seok, Kim Dong-Kyu, Moon Baehun, Cha Seul-Gi, Shin Won-Jung, Kim Sung-Hoon
Department of Anesthesiology and Pain Medicine, BK21 Project, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Republic of Korea.
Cancer Data Science Laboratory, National Cancer Institute, National Institutes of Health, Bethesda, MD, USA.
Comput Methods Programs Biomed. 2025 Sep;269:108871. doi: 10.1016/j.cmpb.2025.108871. Epub 2025 May 21.
Auscultation-based cardiac abnormality detection is valuable screening approach in pediatric populations, particularly in resource-limited settings. However, its clinical utility is often limited by phonocardiogram (PCG) signal variability and a difficulty in distinguishing between pathological and innocent murmurs.
We proposed a framework that leverages temporal convolutional network (TCN)-based feature extraction and information fusion to integrate asynchronously acquired PCG recordings at the patient level. A probabilistic representation of the pathological state was first extracted from segmented PCG signals using a TCN-based model. These segment-level representations were subsequently averaged to generate record- or patient-level features. The framework was designed to accommodate recordings of varying durations and different auscultation locations. Furthermore, we addressed domain adaptation challenges in cardiac abnormality detection by incorporating transfer learning techniques.
The proposed method was evaluated using two large, independent public PCG datasets, demonstrating robust performance at both record and patient levels. While its initial performance on an unseen external dataset was modest, likely due to demographic characteristics and signal acquisition, transfer learning significantly improved the model's performance, yielding an area under the receiver operating characteristic curve of 0.931±0.027 and an area under the precision-recall curve of 0.867±0.064 in external validation. Combining internal and external datasets further enhanced model generalizability.
This proposed framework accommodates multi-channel, variable-length PCG recordings, making it a flexible and accurate solution for detecting pediatric cardiac abnormalities, particularly in low-resource settings. The source code is publicly available on Github (https://github.com/baporlab/pcg_pathological_murmur_detection).
基于听诊的心脏异常检测是儿科人群中有价值的筛查方法,尤其是在资源有限的环境中。然而,其临床效用常常受到心音图(PCG)信号变异性以及区分病理性杂音和生理性杂音困难的限制。
我们提出了一个框架,该框架利用基于时间卷积网络(TCN)的特征提取和信息融合,在患者层面整合异步采集的PCG记录。首先使用基于TCN的模型从分段的PCG信号中提取病理状态的概率表示。随后对这些段级表示进行平均,以生成记录级或患者级特征。该框架旨在适应不同持续时间和不同听诊位置的记录。此外,我们通过纳入迁移学习技术来应对心脏异常检测中的领域适应挑战。
使用两个大型独立的公共PCG数据集对所提出的方法进行了评估,结果表明该方法在记录和患者层面均具有稳健的性能。虽然其在一个未见的外部数据集上的初始性能一般,可能是由于人口统计学特征和信号采集的原因,但迁移学习显著提高了模型的性能,在外部验证中,受试者操作特征曲线下面积为0.931±0.027,精确召回率曲线下面积为0.867±0.064。结合内部和外部数据集进一步增强了模型的泛化能力。
所提出的这个框架适用于多通道、可变长度的PCG记录,使其成为检测儿科心脏异常的灵活且准确的解决方案,尤其是在资源匮乏的环境中。源代码可在Github上公开获取(https://github.com/baporlab/pcg_pathological_murmur_detection)。