Le Thanh-Dung, Jouvet Philippe, Noumeir Rita
Biomedical Information Processing Laboratory, École de Technologie SupérieureUniversity of Quebec Quebec City QC G1K 9H6 Canada.
Interdisciplinary Centre for Security, Reliability, and Trust (SnT)University of Luxembourg 4365 Esch-sur-Alzette Luxembourg.
IEEE J Transl Eng Health Med. 2025 Jun 4;13:261-274. doi: 10.1109/JTEHM.2025.3576570. eCollection 2025.
Transformer-based models have shown outstanding results in natural language processing but face challenges in applications like classifying small-scale clinical texts, especially with constrained computational resources. This study presents a customized Mixture of Expert (MoE) Transformer models for classifying small-scale French clinical texts at CHU Sainte-Justine Hospital. The MoE-Transformer addresses the dual challenges of effective training with limited data and low-resource computation suitable for in-house hospital use. Despite the success of biomedical pre-trained models such as CamemBERT-bio, DrBERT, and AliBERT, their high computational demands make them impractical for many clinical settings. Our MoE-Transformer model not only outperforms DistillBERT, CamemBERT, FlauBERT, and Transformer models on the same dataset but also achieves impressive results: an accuracy of 87%, precision of 87%, recall of 85%, and F1-score of 86%. While the MoE-Transformer does not surpass the performance of biomedical pre-trained BERT models, it can be trained at least 190 times faster, offering a viable alternative for settings with limited data and computational resources. Although the MoE-Transformer addresses challenges of generalization gaps and sharp minima, demonstrating some limitations for efficient and accurate clinical text classification, this model still represents a significant advancement in the field. It is particularly valuable for classifying small French clinical narratives within the privacy and constraints of hospital-based computational resources. Clinical and Translational Impact Statement-This study highlights the potential of customized MoE-Transformers in enhancing clinical text classification, particularly for small-scale datasets like French clinical narratives. The MoE-Transformer's ability to outperform several pre-trained BERT models marks a stride in applying NLP techniques to clinical data and integrating into a Clinical Decision Support System in a Pediatric Intensive Care Unit. The study underscores the importance of model selection and customization in achieving optimal performance for specific clinical applications, especially with limited data availability and within the constraints of hospital-based computational resources.
基于Transformer的模型在自然语言处理中取得了优异的成果,但在对小规模临床文本进行分类等应用中面临挑战,尤其是在计算资源有限的情况下。本研究提出了一种定制的专家混合(MoE)Transformer模型,用于对圣贾斯汀医院的小规模法语临床文本进行分类。MoE-Transformer解决了在有限数据下进行有效训练以及适用于医院内部使用的低资源计算这两个双重挑战。尽管诸如CamemBERT-bio、DrBERT和AliBERT等生物医学预训练模型取得了成功,但它们对计算的高要求使其在许多临床环境中不切实际。我们的MoE-Transformer模型不仅在同一数据集上优于DistillBERT、CamemBERT、FlauBERT和Transformer模型,还取得了令人印象深刻的结果:准确率为87%,精确率为87%,召回率为85%,F1分数为86%。虽然MoE-Transformer没有超过生物医学预训练BERT模型的性能,但它的训练速度至少可以快190倍,为数据和计算资源有限的环境提供了一个可行的替代方案。尽管MoE-Transformer解决了泛化差距和尖锐极小值的挑战,在高效准确的临床文本分类方面表现出一些局限性,但该模型仍然代表了该领域的重大进步。它对于在医院计算资源的隐私和限制范围内对小规模法语临床叙述进行分类特别有价值。临床与转化影响声明——本研究强调了定制的MoE-Transformer在增强临床文本分类方面的潜力,特别是对于像法语临床叙述这样的小规模数据集。MoE-Transformer优于多个预训练BERT模型的能力标志着在将自然语言处理技术应用于临床数据并集成到儿科重症监护病房的临床决策支持系统方面迈出了一大步。该研究强调了模型选择和定制对于实现特定临床应用的最佳性能的重要性,特别是在数据可用性有限以及医院计算资源受限的情况下。