Carrillo-Larco Rodrigo M
Hubert Department of Global Health, Rollins School of Public Health, Emory University, Atlanta, GA, USA.
Pediatr Cardiol. 2024 Jun 27. doi: 10.1007/s00246-024-03561-2.
Research has shown that X-rays and fundus images can classify gender, age group, and race, raising concerns about bias and fairness in medical AI applications. However, the potential for physiological sounds to classify sociodemographic traits has not been investigated. Exploring this gap is crucial for understanding the implications and ensuring fairness in the field of medical sound analysis. We aimed to develop classifiers to determine gender (men/women) based on heart sound recordings and using machine learning (ML). Data-driven ML analysis. We utilized the open-access CirCor DigiScope Phonocardiogram Dataset obtained from cardiac screening programs in Brazil. Volunteers < 21 years of age. Each participant completed a questionnaire and underwent a clinical examination, including electronic auscultation at four cardiac points: aortic (AV), mitral (MV), pulmonary (PV), and tricuspid (TV). We used Mel-frequency cepstral coefficients (MFCCs) to develop the ML classifiers. From each patient and from each auscultation sound recording, we extracted 10 MFCCs. In sensitivity analysis, we additionally extracted 20, 30, 40, and 50 MFCCs. The most effective gender classifier was developed using PV recordings (AUC ROC = 70.3%). The second best came from MV recordings (AUC ROC = 58.8%). AV and TV recordings produced classifiers with an AUC ROC of 56.4% and 56.1%, respectively. Using more MFCCs did not substantially improve the classifiers. It is possible to classify between males and females using phonocardiogram data. As health-related audio recordings become more prominent in ML applications, research is required to explore if these recordings contain signals that could distinguish sociodemographic features.
研究表明,X射线和眼底图像可以对性别、年龄组和种族进行分类,这引发了人们对医学人工智能应用中偏见和公平性的担忧。然而,生理声音对社会人口特征进行分类的潜力尚未得到研究。探索这一差距对于理解医学声音分析领域的影响并确保公平性至关重要。我们旨在开发基于心音记录并使用机器学习(ML)来确定性别(男性/女性)的分类器。数据驱动的ML分析。我们利用了从巴西心脏筛查项目中获得的开放获取的CirCor DigiScope心音图数据集。志愿者年龄小于21岁。每位参与者都完成了一份问卷并接受了临床检查,包括在四个心脏点进行电子听诊:主动脉(AV)、二尖瓣(MV)、肺动脉(PV)和三尖瓣(TV)。我们使用梅尔频率倒谱系数(MFCCs)来开发ML分类器。从每位患者和每次听诊录音中,我们提取了10个MFCCs。在敏感性分析中,我们还额外提取了20、30、40和50个MFCCs。最有效的性别分类器是使用PV录音开发的(AUC ROC = 70.3%)。第二好的来自MV录音(AUC ROC = 58.8%)。AV和TV录音产生的分类器的AUC ROC分别为56.4%和56.1%。使用更多的MFCCs并没有显著改善分类器。使用心音图数据可以对男性和女性进行分类。随着与健康相关的音频记录在ML应用中变得更加突出,需要进行研究以探索这些记录是否包含可以区分社会人口特征的信号。