EIHW - Chair of Embedded Intelligence for Health Care and Wellbeing, University of Augsburg, Augsburg, Germany; L3S Research Center, Hannover, Germany.
GLAM - Group on Language, Audio, & Music, Imperial College London, London, United Kingdom.
J Voice. 2024 Nov;38(6):1264-1277. doi: 10.1016/j.jvoice.2022.06.011. Epub 2022 Jun 15.
The coronavirus disease 2019 (COVID-19) has caused a crisis worldwide. Amounts of efforts have been made to prevent and control COVID-19's transmission, from early screenings to vaccinations and treatments. Recently, due to the spring up of many automatic disease recognition applications based on machine listening techniques, it would be fast and cheap to detect COVID-19 from recordings of cough, a key symptom of COVID-19. To date, knowledge of the acoustic characteristics of COVID-19 cough sounds is limited but would be essential for structuring effective and robust machine learning models. The present study aims to explore acoustic features for distinguishing COVID-19 positive individuals from COVID-19 negative ones based on their cough sounds.
By applying conventional inferential statistics, we analyze the acoustic correlates of COVID-19 cough sounds based on the ComParE feature set, i.e., a standardized set of 6,373 acoustic higher-level features. Furthermore, we train automatic COVID-19 detection models with machine learning methods and explore the latent features by evaluating the contribution of all features to the COVID-19 status predictions.
The experimental results demonstrate that a set of acoustic parameters of cough sounds, e.g., statistical functionals of the root mean square energy and Mel-frequency cepstral coefficients, bear essential acoustic information in terms of effect sizes for the differentiation between COVID-19 positive and COVID-19 negative cough samples. Our general automatic COVID-19 detection model performs significantly above chance level, i.e., at an unweighted average recall (UAR) of 0.632, on a data set consisting of 1,411 cough samples (COVID-19 positive/negative: 210/1,201).
Based on the acoustic correlates analysis on the ComParE feature set and the feature analysis in the effective COVID-19 detection approach, we find that several acoustic features that show higher effects in conventional group difference testing are also higher weighted in the machine learning models.
新型冠状病毒病(COVID-19)在全球范围内造成了危机。人们已经做出了大量努力来预防和控制 COVID-19 的传播,从早期筛查到疫苗接种和治疗。最近,由于许多基于机器监听技术的自动疾病识别应用程序的兴起,从 COVID-19 的关键症状咳嗽的录音中快速且廉价地检测 COVID-19 将变得更加容易。迄今为止,有关 COVID-19 咳嗽声音的声学特征的知识有限,但对于构建有效的机器学习模型至关重要。本研究旨在探索基于咳嗽声音区分 COVID-19 阳性和 COVID-19 阴性个体的声学特征。
通过应用常规推理统计,我们基于 ComParE 特征集(即标准化的 6373 个声学高级特征集)分析 COVID-19 咳嗽声音的声学相关性。此外,我们使用机器学习方法训练自动 COVID-19 检测模型,并通过评估所有特征对 COVID-19 状态预测的贡献来探索潜在特征。
实验结果表明,咳嗽声音的一组声学参数,例如均方根能量和梅尔频率倒谱系数的统计函数,在区分 COVID-19 阳性和 COVID-19 阴性咳嗽样本方面具有重要的声学信息。我们的通用自动 COVID-19 检测模型在一个包含 1411 个咳嗽样本(COVID-19 阳性/阴性:210/1201)的数据集中的表现明显优于机会水平,即未加权平均召回率(UAR)为 0.632。
基于 ComParE 特征集的声学相关性分析和有效 COVID-19 检测方法中的特征分析,我们发现,在常规组差异测试中显示更高效果的一些声学特征在机器学习模型中也具有更高的权重。