Tran Bich Anh, Dao Thao Thi Phuong, Dung Ho Dang Quy, Van Ngoc Boi, Ha Chanh Cong, Pham Nam Hoang, Nguyen Tu Cong Huyen Ton Nu Cam, Nguyen Tan-Cong, Pham Minh-Khoi, Tran Mai-Khiem, Tran Truong Minh, Tran Minh-Triet
Otorhinolaryngology Department, Cho Ray Hospital, Ho Chi Minh City, Viet Nam.
University of Science, VNUHCM, Ho Chi Minh City, Viet Nam; John von Neumann Institute, VNUHCM, Ho Chi Minh City, Viet Nam; Vietnam National University, Ho Chi Minh City, Viet Nam; Department of Otolaryngology, Thong Nhat Hospital, Ho Chi Minh City, Viet Nam.
Am J Otolaryngol. 2023 May-Jun;44(3):103800. doi: 10.1016/j.amjoto.2023.103800. Epub 2023 Feb 24.
To collect a dataset with adequate laryngoscopy images and identify the appearance of vocal folds and their lesions in flexible laryngoscopy images by objective deep learning models.
We adopted a number of novel deep learning models to train and classify 4549 flexible laryngoscopy images as no vocal fold, normal vocal folds, and abnormal vocal folds. This could help these models recognize vocal folds and their lesions within these images. Ultimately, we made a comparison between the results of the state-of-the-art deep learning models, and another comparison of the results between the computer-aided classification system and ENT doctors.
This study exhibited the performance of the deep learning models by evaluating laryngoscopy images collected from 876 patients. The efficiency of the Xception model was higher and steadier than almost the rest of the models. The accuracy of no vocal fold, normal vocal folds, and vocal fold abnormalities on this model were 98.90 %, 97.36 %, and 96.26 %, respectively. Compared to our ENT doctors, the Xception model produced better results than a junior doctor and was near an expert.
Our results show that current deep learning models can classify vocal fold images well and effectively assist physicians in vocal fold identification and classification of normal or abnormal vocal folds.
收集包含足够喉镜图像的数据集,并通过客观深度学习模型识别柔性喉镜图像中声带的外观及其病变。
我们采用了多种新型深度学习模型,对4549张柔性喉镜图像进行训练和分类,分为无声带、正常声带和异常声带。这有助于这些模型识别这些图像中的声带及其病变。最终,我们对最先进的深度学习模型的结果进行了比较,还对计算机辅助分类系统与耳鼻喉科医生的结果进行了另一项比较。
本研究通过评估从876名患者收集的喉镜图像,展示了深度学习模型的性能。Xception模型的效率比几乎其他所有模型都更高且更稳定。该模型上无声带、正常声带和声带异常的准确率分别为98.90%、97.36%和96.26%。与我们的耳鼻喉科医生相比,Xception模型的结果比初级医生更好,接近专家水平。
我们的结果表明,当前的深度学习模型可以很好地对声带图像进行分类,并有效地协助医生进行声带识别以及正常或异常声带的分类。