Zheng Ziyu, Yang Xiaying, Wu Shengjie, Zhang Shijie, Lyu Guorong, Liu Peizhong, Wang Jun, He Shaozheng
Department of Information Construction and Management.
Department of Ultrasound, Second Affiliated Hospital of Fujian Medical University, Quanzhou 362000, China.
Nan Fang Yi Ke Da Xue Xue Bao. 2025 Jul 20;45(7):1563-1570. doi: 10.12122/j.issn.1673-4254.2025.07.24.
To construct an intelligent analysis model for classifying fetal orientation during intrapartum ultrasound videos based on multi-feature fusion.
The proposed model consists of the Input, Backbone Network and Classification Head modules. The Input module carries out data augmentation to improve the sample quality and generalization ability of the model. The Backbone Network was responsible for feature extraction based on Yolov8 combined with CBAM, ECA, PSA attention mechanism and AIFI feature interaction module. The Classification Head consists of a convolutional layer and a softmax function to output the final probability value of each class. The images of the key structures (the eyes, face, head, thalamus, and spine) were annotated with frames by physicians for model training to improve the classification accuracy of the anterior occipital, posterior occipital, and transverse occipital orientations.
The experimental results showed that the proposed model had excellent performance in the tire orientation classification task with the classification accuracy reaching 0.984, an area under the PR curve (average accuracy) of 0.993, and area under the ROC curve of 0.984, and a kappa consistency test score of 0.974. The prediction results by the deep learning model were highly consistent with the actual classification results.
The multi-feature fusion model proposed in this study can efficiently and accurately classify fetal orientation in intrapartum ultrasound videos.
构建基于多特征融合的产时超声视频中胎儿方位分类智能分析模型。
所提出的模型由输入模块、骨干网络和分类头模块组成。输入模块进行数据增强以提高模型的样本质量和泛化能力。骨干网络负责基于结合了CBAM、ECA、PSA注意力机制和AIFI特征交互模块的Yolov8进行特征提取。分类头由一个卷积层和一个softmax函数组成,用于输出每个类别的最终概率值。由医生对关键结构(眼睛、面部、头部、丘脑和脊柱)的图像进行逐帧标注以用于模型训练,从而提高枕前位、枕后位和枕横位的分类准确率。
实验结果表明,所提出的模型在胎儿方位分类任务中具有优异的性能,分类准确率达到0.984,PR曲线下面积(平均准确率)为0.993,ROC曲线下面积为0.984,kappa一致性检验得分0.974。深度学习模型的预测结果与实际分类结果高度一致。
本研究提出的多特征融合模型能够高效、准确地对产时超声视频中的胎儿方位进行分类。