Computer Science Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia.
Computer Science Department, College of Computer and Information Sciences, Al-Imam Mohammad Ibn Saud Islamic University, Riyadh 11432, Saudi Arabia.
Sensors (Basel). 2023 Sep 15;23(18):7924. doi: 10.3390/s23187924.
Scoring polysomnography for obstructive sleep apnea diagnosis is a laborious, long, and costly process. Machine learning approaches, such as deep neural networks, can reduce scoring time and costs. However, most methods require prior filtering and preprocessing of the raw signal. Our work presents a novel method for diagnosing obstructive sleep apnea using a transformer neural network with learnable positional encoding, which outperforms existing state-of-the-art solutions. This approach has the potential to improve the diagnostic performance of oximetry for obstructive sleep apnea and reduce the time and costs associated with traditional polysomnography. Contrary to existing approaches, our approach performs annotations at one-second granularity. Allowing physicians to interpret the model's outcome. In addition, we tested different positional encoding designs as the first layer of the model, and the best results were achieved using a learnable positional encoding based on an autoencoder with structural novelty. In addition, we tried different temporal resolutions with various granularity levels from 1 to 360 s. All experiments were carried out on an independent test set from the public OSASUD dataset and showed that our approach outperforms current state-of-the-art solutions with a satisfactory AUC of 0.89, accuracy of 0.80, and F1-score of 0.79.
阻塞性睡眠呼吸暂停诊断的多导睡眠图评分是一项费力、冗长且昂贵的过程。机器学习方法,如深度神经网络,可以减少评分时间和成本。然而,大多数方法需要对原始信号进行预先过滤和预处理。我们的工作提出了一种使用具有可学习位置编码的变压器神经网络诊断阻塞性睡眠呼吸暂停的新方法,该方法优于现有的最先进的解决方案。这种方法有可能提高血氧仪诊断阻塞性睡眠呼吸暂停的性能,并降低与传统多导睡眠图相关的时间和成本。与现有方法不同,我们的方法以一秒为粒度进行注释。允许医生解释模型的结果。此外,我们还测试了模型第一层的不同位置编码设计,使用基于自动编码器的可学习位置编码取得了最佳效果,该自动编码器具有新颖的结构。此外,我们尝试了不同的时间分辨率和各种粒度级别,从 1 秒到 360 秒不等。所有实验都是在公共 OSASUD 数据集的独立测试集上进行的,结果表明我们的方法优于当前最先进的解决方案,AUC 为 0.89,准确率为 0.80,F1 得分为 0.79。