Supratak Akara, Haddawy Peter
Faculty of ICT, Mahidol University, 999 Phuttamonthon 4 Road, Nakhon Pathom, 73170, Thailand.
Faculty of ICT, Mahidol University, 999 Phuttamonthon 4 Road, Nakhon Pathom, 73170, Thailand; Bremen Spatial Cognition Center, University of Bremen, Bremen, Germany.
Artif Intell Med. 2023 May;139:102540. doi: 10.1016/j.artmed.2023.102540. Epub 2023 Mar 31.
Deep learning models for scoring sleep stages based on single-channel EEG have been proposed as a promising method for remote sleep monitoring. However, applying these models to new datasets, particularly from wearable devices, raises two questions. First, when annotations on a target dataset are unavailable, which different data characteristics affect the sleep stage scoring performance the most and by how much? Second, when annotations are available, which dataset should be used as the source of transfer learning to optimize performance? In this paper, we propose a novel method for computationally quantifying the impact of different data characteristics on the transferability of deep learning models. Quantification is accomplished by training and evaluating two models with significant architectural differences, TinySleepNet and U-Time, under various transfer configurations in which the source and target datasets have different recording channels, recording environments, and subject conditions. For the first question, the environment had the highest impact on sleep stage scoring performance, with performance degrading by over 14% when sleep annotations were unavailable. For the second question, the most useful transfer sources for TinySleepNet and the U-Time models were MASS-SS1 and ISRUC-SG1, containing a high percentage of N1 (the rarest sleep stage) relative to the others. The frontal and central EEGs were preferred for TinySleepNet. The proposed approach enables full utilization of existing sleep datasets for training and planning model transfer to maximize the sleep stage scoring performance on a target problem when sleep annotations are limited or unavailable, supporting the realization of remote sleep monitoring.
基于单通道脑电图对睡眠阶段进行评分的深度学习模型已被提出,作为一种有前景的远程睡眠监测方法。然而,将这些模型应用于新数据集,特别是来自可穿戴设备的数据集,会引发两个问题。首先,当目标数据集没有注释时,哪些不同的数据特征对睡眠阶段评分性能影响最大,影响程度有多大?其次,当有注释时,应该使用哪个数据集作为迁移学习的源来优化性能?在本文中,我们提出了一种新颖的方法,用于通过计算量化不同数据特征对深度学习模型可迁移性的影响。量化是通过在各种迁移配置下训练和评估两个具有显著架构差异的模型TinySleepNet和U-Time来完成的,在这些配置中,源数据集和目标数据集具有不同的记录通道、记录环境和受试者条件。对于第一个问题,环境对睡眠阶段评分性能的影响最大,当没有睡眠注释时,性能下降超过14%。对于第二个问题,TinySleepNet和U-Time模型最有用的迁移源是MASS-SS1和ISRUC-SG1,相对于其他数据集,它们包含较高百分比的N1(最罕见的睡眠阶段)。TinySleepNet更倾向于使用额叶和中央脑电图。当睡眠注释有限或不可用时,所提出的方法能够充分利用现有的睡眠数据集进行训练和规划模型迁移,以最大化目标问题上的睡眠阶段评分性能,支持远程睡眠监测的实现。