Hernandez-Torres Sofia I, Hennessey Ryan P, Snider Eric J
U.S. Army Institute of Surgical Research, JBSA Fort Sam Houston, San Antonio, TX 78234, USA.
Bioengineering (Basel). 2023 Jul 5;10(7):807. doi: 10.3390/bioengineering10070807.
Ultrasound imaging is a critical tool for triaging and diagnosing subjects but only if images can be properly interpreted. Unfortunately, in remote or military medicine situations, the expertise to interpret images can be lacking. Machine-learning image interpretation models that are explainable to the end user and deployable in real time with ultrasound equipment have the potential to solve this problem. We have previously shown how a YOLOv3 (You Only Look Once) object detection algorithm can be used for tracking shrapnel, artery, vein, and nerve fiber bundle features in a tissue phantom. However, real-time implementation of an object detection model requires optimizing model inference time. Here, we compare the performance of five different object detection deep-learning models with varying architectures and trainable parameters to determine which model is most suitable for this shrapnel-tracking ultrasound image application. We used a dataset of more than 16,000 ultrasound images from gelatin tissue phantoms containing artery, vein, nerve fiber, and shrapnel features for training and evaluating each model. Every object detection model surpassed 0.85 mean average precision except for the detection transformer model. Overall, the YOLOv7tiny model had the higher mean average precision and quickest inference time, making it the obvious model choice for this ultrasound imaging application. Other object detection models were overfitting the data as was determined by lower testing performance compared with higher training performance. In summary, the YOLOv7tiny object detection model had the best mean average precision and inference time and was selected as optimal for this application. Next steps will implement this object detection algorithm for real-time applications, an important next step in translating AI models for emergency and military medicine.
超声成像对于对受试者进行分类和诊断而言是一项关键工具,但前提是图像能够得到正确解读。不幸的是,在偏远地区或军事医疗场景中,可能缺乏解读图像的专业知识。能够向终端用户作出解释并可与超声设备实时部署的机器学习图像解读模型有潜力解决这一问题。我们之前已经展示了如何将YOLOv3(你只看一次)目标检测算法用于在组织模型中追踪弹片、动脉、静脉和神经纤维束特征。然而,目标检测模型的实时实现需要优化模型推理时间。在此,我们比较了五种具有不同架构和可训练参数的不同目标检测深度学习模型的性能,以确定哪种模型最适合这种弹片追踪超声图像应用。我们使用了一个包含动脉、静脉、神经纤维和弹片特征的明胶组织模型的16000多张超声图像数据集来训练和评估每个模型。除了检测变压器模型外,每个目标检测模型的平均精度均超过0.85。总体而言,YOLOv7tiny模型具有更高的平均精度和最快的推理时间,使其成为这种超声成像应用的明显模型选择。与较高的训练性能相比,其他目标检测模型的测试性能较低,表明它们对数据存在过拟合现象。总之,YOLOv7tiny目标检测模型具有最佳的平均精度和推理时间,并被选为该应用的最佳模型。下一步将把这种目标检测算法应用于实时场景,这是将人工智能模型转化用于急救和军事医疗的重要后续步骤。