Yang Jun, Wang Qichen, Dong Xiaoxing, Shen Tao
Faculty of Information Engineering and Automation, Kunming University of Science and Technology, No.727 Jingming South Road, Kunming, 650504 Yunnan China.
First People's Hospital of Yunnan Province, No.157 Jinbi Road, Kunming, 650032 Yunnan China.
Health Inf Sci Syst. 2025 Jan 10;13(1):15. doi: 10.1007/s13755-024-00328-0. eCollection 2025 Dec.
For diagnosing mental health conditions and assessing sleep quality, the classification of sleep stages is essential. Although deep learning-based methods are effective in this field, they often fail to capture sufficient features or adequately synthesize information from various sources. For the purpose of improving the accuracy of sleep stage classification, our methodology includes extracting a diverse array of features from polysomnography signals, along with their transformed graph and time-frequency representations. We have developed specific feature extraction modules tailored for each distinct view. To efficiently integrate and categorize the features derived from these different perspectives, we propose a cross-attention fusion mechanism. This mechanism is designed to adaptively merge complex sleep features, facilitating a more robust classification process. More specifically, our strategy includes the development of an efficient fusion network with multi-view features for classifying sleep stages that incorporates brain connectivity and combines both temporal and spectral elements for sleep stage analysis. This network employs a systematic approach to extract spatio-temporal-frequency features and uses cross-attention to merge features from different views effectively. In the experiments we conducted on the ISRUC public datasets, we found that our approach outperformed other proposed methods. In the ablation experiments, there was also a 2% improvement over the baseline model. Our research indicates that multi-view feature fusion methods with a cross-attention mechanism have strong potential in sleep stage classification.
对于诊断心理健康状况和评估睡眠质量而言,睡眠阶段的分类至关重要。尽管基于深度学习的方法在该领域有效,但它们往往无法捕捉到足够的特征或充分综合来自各种来源的信息。为了提高睡眠阶段分类的准确性,我们的方法包括从多导睡眠图信号中提取各种各样的特征,以及它们的变换后的图形和时频表示。我们针对每个不同的视图开发了特定的特征提取模块。为了有效地整合和分类从这些不同视角派生的特征,我们提出了一种交叉注意力融合机制。该机制旨在自适应地合并复杂的睡眠特征,促进更稳健的分类过程。更具体地说,我们的策略包括开发一个具有多视图特征的高效融合网络,用于对睡眠阶段进行分类,该网络纳入了大脑连通性,并结合了时间和频谱元素进行睡眠阶段分析。该网络采用系统的方法来提取时空频率特征,并使用交叉注意力有效地合并来自不同视图的特征。在我们对ISRUC公共数据集进行的实验中,我们发现我们的方法优于其他提出的方法。在消融实验中,与基线模型相比也有2%的提升。我们的研究表明,具有交叉注意力机制的多视图特征融合方法在睡眠阶段分类中具有强大的潜力。