Key Laboratory of Optoelectronic Technology and Systems of Ministry of Education, College of Optoelectronic Engineering, Chongqing University, Campus A, Room 1105 of Main Building, 174 Shazhen Street, Shapinba District, Chongqing, 400040, China.
College of Computer and Information Science, Chongqing Normal University, Chongqing, 400050, China.
J Digit Imaging. 2019 Apr;32(2):336-348. doi: 10.1007/s10278-018-0140-5.
Automatic vertebrae localization and identification in medical computed tomography (CT) scans is of great value for computer-aided spine diseases diagnosis. In order to overcome the disadvantages of the approaches employing hand-crafted, low-level features and based on field-of-view priori assumption of spine structure, an automatic method is proposed to localize and identify vertebrae by combining deep stacked sparse autoencoder (SSAE) contextual features and structured regression forest (SRF). The method employs SSAE to learn image deep contextual features instead of hand-crafted ones by building larger-range input samples to improve their contextual discrimination ability. In the localization and identification stage, it incorporates the SRF model to achieve whole spine localization, then screens those vertebrae within the image, thus relieves the assumption that the part of spine in the field of image is visible. In the end, the output distribution of SRF and spine CT scans properties are assembled to develop a two-stage progressive refining strategy, where the mean-shift kernel density estimation and Otsu method instead of Markov random field (MRF) are adopted to reduce model complexity and refine vertebrae localization results. Extensive evaluation was performed on a challenging data set of 98 spine CT scans. Compared with the hidden Markov model and the method based on convolutional neural network (CNN), the proposed approach could effectively and automatically locate and identify spinal targets in CT scans, and achieve higher localization accuracy, low model complexity, and no need for any assumptions about visual field in CT scans.
医学计算机断层扫描(CT)中自动脊椎定位与识别对于计算机辅助脊柱疾病诊断具有重要价值。为了克服基于手工制作的低级特征和脊柱结构视场先验假设的方法的缺点,提出了一种通过结合深度堆叠稀疏自编码器(SSAE)上下文特征和结构化回归森林(SRF)来自动定位和识别脊椎的方法。该方法通过构建更大范围的输入样本来学习图像的深度上下文特征,而不是使用手工制作的特征,从而提高其上下文辨别能力。在定位和识别阶段,它结合了 SRF 模型来实现整个脊柱的定位,然后筛选图像内的那些脊椎,从而减轻了图像视场中部分脊柱可见的假设。最后,将 SRF 的输出分布和脊椎 CT 扫描特性组合起来,开发了一种两阶段渐进细化策略,其中采用均值漂移核密度估计和 Otsu 方法而不是马尔可夫随机场(MRF)来降低模型复杂度并细化脊椎定位结果。在一个具有挑战性的 98 个脊椎 CT 扫描数据集上进行了广泛的评估。与隐马尔可夫模型和基于卷积神经网络(CNN)的方法相比,所提出的方法可以有效地自动定位和识别 CT 扫描中的脊柱目标,并且具有更高的定位精度、低模型复杂度,并且不需要对 CT 扫描中的视场进行任何假设。