Nguyen Cuong V, Do Cuong D
College of Engineering and Computer Science, VinUniversity, Hanoi, Vietnam.
VinUni-Illinois Smart Health Center, VinUniversity, Hanoi, Vietnam.
PLoS One. 2025 May 19;20(5):e0316043. doi: 10.1371/journal.pone.0316043. eCollection 2025.
The adoption of deep learning in ECG diagnosis is often hindered by the scarcity of large, well-labeled datasets in real-world scenarios, leading to the use of transfer learning to leverage features learned from larger datasets. Yet the prevailing assumption that transfer learning consistently outperforms training from scratch has never been systematically validated. In this study, we conduct the first extensive empirical study on the effectiveness of transfer learning in multi-label ECG classification, by investigating comparing the fine-tuning performance with that of training from scratch, covering a variety of ECG datasets and deep neural networks. Firstly, We confirm that fine-tuning is the preferable choice for small downstream datasets; however, it does not necessarily improve performance. Secondly, the improvement from fine-tuning declines when the downstream dataset grows. With a sufficiently large dataset, training from scratch can achieve comparable performance, albeit requiring a longer training time to catch up. Thirdly, fine-tuning can accelerate convergence, resulting in faster training process and lower computing cost. Finally, we find that transfer learning exhibits better compatibility with convolutional neural networks than with recurrent neural networks, which are the two most prevalent architectures for time-series ECG applications. Our results underscore the importance of transfer learning in ECG diagnosis, yet depending on the amount of available data, researchers may opt not to use it, considering the non-negligible cost associated with pre-training.
在现实场景中,深度学习在心电图诊断中的应用常常受到大型、标注良好的数据集稀缺的阻碍,这导致人们使用迁移学习来利用从更大数据集中学到的特征。然而,迁移学习始终优于从头开始训练这一普遍假设从未得到系统验证。在本研究中,我们通过研究比较微调性能与从头开始训练的性能,对迁移学习在多标签心电图分类中的有效性进行了首次广泛的实证研究,涵盖了各种心电图数据集和深度神经网络。首先,我们证实微调是小型下游数据集的首选;然而,它并不一定会提高性能。其次,当下游数据集增大时,微调带来的性能提升会下降。对于足够大的数据集,从头开始训练可以达到可比的性能,尽管需要更长的训练时间才能赶上。第三,微调可以加速收敛,从而实现更快的训练过程和更低的计算成本。最后,我们发现迁移学习与卷积神经网络的兼容性优于与循环神经网络的兼容性,这两种网络是时间序列心电图应用中最普遍的两种架构。我们的结果强调了迁移学习在心电图诊断中的重要性,但考虑到预训练相关的不可忽视的成本,根据可用数据量,研究人员可能会选择不使用它。