Supratak Akara, Guo Yike
Annu Int Conf IEEE Eng Med Biol Soc. 2020 Jul;2020:641-644. doi: 10.1109/EMBC44109.2020.9176741.
Deep learning has become popular for automatic sleep stage scoring due to its capability to extract useful features from raw signals. Most of the existing models, however, have been overengineered to consist of many layers or have introduced additional steps in the processing pipeline, such as converting signals to spectrogram-based images. They require to be trained on a large dataset to prevent the overfitting problem (but most of the sleep datasets contain a limited amount of class-imbalanced data) and are difficult to be applied (as there are many hyperparameters to be configured in the pipeline). In this paper, we propose an efficient deep learning model, named TinySleepNet, and a novel technique to effectively train the model end-to-end for automatic sleep stage scoring based on raw single-channel EEG. Our model consists of a less number of model parameters to be trained compared to the existing ones, requiring a less amount of training data and computational resources. Our training technique incorporates data augmentation that can make our model be more robust the shift along the time axis, and can prevent the model from remembering the sequence of sleep stages. We evaluated our model on seven public sleep datasets that have different characteristics in terms of scoring criteria and recording channels and environments. The results show that, with the same model architecture and the training parameters, our method achieves a similar (or better) performance compared to the state-of-the-art methods on all datasets. This demonstrates that our method can generalize well to the largest number of different datasets.
由于深度学习能够从原始信号中提取有用特征,它已在自动睡眠阶段评分中变得流行起来。然而,大多数现有模型都设计过度,包含许多层,或者在处理流程中引入了额外步骤,例如将信号转换为基于频谱图的图像。它们需要在大型数据集上进行训练以防止过拟合问题(但大多数睡眠数据集包含有限数量的类别不平衡数据),并且难以应用(因为在流程中有许多超参数需要配置)。在本文中,我们提出了一种高效的深度学习模型,名为TinySleepNet,以及一种新颖的技术,用于基于原始单通道脑电图对模型进行端到端的有效训练,以实现自动睡眠阶段评分。与现有模型相比,我们的模型需要训练的模型参数数量更少,所需的训练数据和计算资源也更少。我们的训练技术纳入了数据增强,这可以使我们的模型对沿时间轴的偏移更具鲁棒性,并能防止模型记住睡眠阶段的顺序。我们在七个公共睡眠数据集上评估了我们的模型,这些数据集在评分标准、记录通道和环境方面具有不同的特征。结果表明,在相同的模型架构和训练参数下,我们的方法在所有数据集上与最先进的方法相比都取得了相似(或更好)的性能。这表明我们的方法可以很好地推广到最多数量的不同数据集。