Wang Ganhong, Zhang Zihao, Xia Kaijian, Zhou Yanting, Xi Meijuan, Chen Jian
Department of Gastroenterology, Changshu Hospital of TCM, Changshu 215500, Jiangsu Province, China.
Shanghai Haoxiong Educational Technology Co., Ltd.
Zhongguo Zhen Jiu. 2025 Apr 12;45(4):413-420. doi: 10.13703/j.0255-2930.20240611-0001. Epub 2025 Jan 7.
To develop an artificial intelligence-assisted system for the automatic detection of the features of common 21 auricular points based on the YOLOv8 neural network.
A total of 660 human auricular images from three research centers were collected from June 2019 to February 2024. The rectangle boxes and features of images were annotated using the LabelMe5.3.1 tool and converted them into a format compatible with the YOLO model. Using these data, transfer learning and fine-tuning training were conducted on different scales of pretrained YOLO neural network models. The model's performance was evaluated on validation and test sets, including the mean average precision (mAP) at various thresholds, recall rate (recall), frames per second (FPS) and confusion matrices. Finally, the model was deployed on a local computer, and the real-time detection of human auricular images was conducted using a camera.
Five different versions of the YOLOv8 key-point detection model were developed, including YOLOv8n, YOLOv8s, YOLOv8m, YOLOv8l, and YOLOv8x. On the validation set, YOLOv8n showed the best performance in terms of speed (225.736 frames per second) and precision (0.998). On the external test set, YOLOv8n achieved the accuracy of 0.991, the sensitivity of 1.0, and the F1 score of 0.995. The localization performance of auricular point features showed the average accuracy of 0.990, the precision of 0.995, and the recall of 0.997 under 50% intersection ration (mAP).
The key-point detection model of 21 common auricular points based on YOLOv8n exhibits the excellent predictive performance, which is capable of rapidly and automatically locating and classifying auricular points.
基于YOLOv8神经网络开发一种用于自动检测21个常见耳穴特征的人工智能辅助系统。
2019年6月至2024年2月,从三个研究中心收集了660张人体耳部图像。使用LabelMe5.3.1工具对图像的矩形框和特征进行标注,并将其转换为与YOLO模型兼容的格式。利用这些数据,对不同尺度的预训练YOLO神经网络模型进行迁移学习和微调训练。在验证集和测试集上评估模型性能,包括不同阈值下的平均精度均值(mAP)、召回率(recall)、每秒帧数(FPS)和混淆矩阵。最后,将模型部署在本地计算机上,使用摄像头对人体耳部图像进行实时检测。
开发了五个不同版本的YOLOv8关键点检测模型,包括YOLOv8n、YOLOv8s、YOLOv8m、YOLOv8l和YOLOv8x。在验证集上,YOLOv8n在速度(每秒225.736帧)和精度(0.998)方面表现最佳。在外部测试集上,YOLOv8n的准确率为0.991,灵敏度为1.0,F1分数为0.995。耳穴特征的定位性能在50%交并比(mAP)下显示平均准确率为0.990,精度为0.995,召回率为0.997。
基于YOLOv8n的21个常见耳穴关键点检测模型具有优异的预测性能,能够快速自动定位和分类耳穴。