Lu Hongjiang, Liu Miao, Yu Kun, Fang Yuan, Zhao Jing, Shi Yang
Department of Radiology, The 903rd Hospital of PLA Joint Logistics Support Force (Xihu Hospital Affiliated with Hangzhou Medical College), Hangzhou, Zhejiang, China.
Department of Head and Neck Surgery, Zhejiang Provincial People's Hospital (Affiliated People's Hospital, Hangzhou Medical College), Hangzhou, Zhejiang, China.
Br J Hosp Med (Lond). 2025 Sep 25;86(9):1-22. doi: 10.12968/hmed.2025.0443. Epub 2025 Sep 24.
Spinal disorders, such as herniated discs and scoliosis, are highly prevalent conditions with rising incidence in the aging global population. Accurate analysis of spinal anatomical structures is a critical prerequisite for achieving high-precision positioning with surgical navigation robots. However, traditional manual segmentation methods are limited by issues such as low efficiency and poor consistency. This work aims to develop a fully automated deep learning-based vertebral segmentation and labeling workflow to provide efficient and accurate preoperative analysis support for spine surgery navigation robots. In the localization stage, the You Only Look Once version 7 (YOLOv7) network was utilized to predict the bounding boxes of individual vertebrae on computed tomography (CT) sagittal slices, transforming the 3D localization problem into a 2D one. Subsequently, the Density-Based Spatial Clustering of Applications with Noise (DBSCAN) clustering algorithm was employed to aggregate the 2D detection results into 3D vertebral centers. This approach significantly reduces inference time and enhances localization accuracy. In the segmentation stage, a 3D U-Net model integrated with an attention mechanism was trained using the region of interest (ROI) based on the vertebral center as input, effectively extracting the 3D structural features of vertebrae to achieve precise segmentation. In the labeling stage, a vertebra labeling network was trained by combining deep learning architectures-ResNet and Transformer, which are capable of extracting rich intervertebral features, to obtain the final labeling results through post-processing based on positional logic analysis. To verify the effectiveness of this workflow, experiments were conducted on a dataset comprising 106 spinal CT datasets sourced from various devices, covering a wide range of clinical scenarios. The results demonstrate that the method performed excellently in the three key tasks of localization, segmentation, and labeling, with a Mean Localization Error (MLE) of 1.42 mm. The segmentation accuracy metrics included a Dice Similarity Coefficient (DSC) of 0.968 ± 0.014, Intersection over Union (IoU) of 0.879 ± 0.018, Pixel Accuracy (PA) of 0.988 ± 0.005, mean symmetric distance (MSD) of 1.09 ± 0.19 mm, and Hausdorff Distance (HD) of 5.42 ± 2.05 mm. The degree of classification accuracy reached up to 94.36%. These quantitative assessments and visualizations confirm the effectiveness of our method (vertebra localization, vertebra segmentation and vertebra labeling), indicating its potential for deployment in spinal surgery navigation robots to provide accurate and efficient preoperative analysis and navigation support for spinal surgeries.
脊柱疾病,如椎间盘突出症和脊柱侧弯,在全球老龄化人口中极为普遍且发病率不断上升。准确分析脊柱解剖结构是手术导航机器人实现高精度定位的关键前提。然而,传统的手动分割方法存在效率低和一致性差等问题。这项工作旨在开发一种基于深度学习的全自动椎体分割和标记工作流程,为脊柱手术导航机器人提供高效、准确的术前分析支持。在定位阶段,利用You Only Look Once版本7(YOLOv7)网络预测计算机断层扫描(CT)矢状面上单个椎体的边界框,将三维定位问题转化为二维问题。随后,采用基于密度的带噪声空间聚类(DBSCAN)算法将二维检测结果聚合为三维椎体中心。这种方法显著减少了推理时间并提高了定位精度。在分割阶段,使用基于椎体中心的感兴趣区域(ROI)训练一个集成了注意力机制的三维U-Net模型,有效提取椎体的三维结构特征以实现精确分割。在标记阶段,通过结合能够提取丰富椎间特征的深度学习架构ResNet和Transformer训练一个椎体标记网络,通过基于位置逻辑分析的后处理获得最终标记结果。为验证该工作流程的有效性,在一个包含106个来自各种设备的脊柱CT数据集的数据集上进行了实验,涵盖了广泛的临床场景。结果表明,该方法在定位、分割和标记这三个关键任务中表现出色,平均定位误差(MLE)为1.42毫米。分割精度指标包括骰子相似系数(DSC)为0.968±0.014、交并比(IoU)为0.879±0.018、像素精度(PA)为0.988±0.005、平均对称距离(MSD)为1.09±0.19毫米和豪斯多夫距离(HD)为5.42±2.05毫米。分类准确率高达94.36%。这些定量评估和可视化结果证实了我们方法(椎体定位、椎体分割和椎体标记)的有效性,表明其在脊柱手术导航机器人中部署的潜力,可为脊柱手术提供准确、高效的术前分析和导航支持。