Department of Radiology, University of Pittsburgh Medical Center (UPMC), 200 Lothrop St., Pittsburgh, PA, 15213, USA.
Center for Research Computing, University of Pittsburgh, Pittsburgh, PA, 15260, USA.
J Digit Imaging. 2019 Aug;32(4):672-677. doi: 10.1007/s10278-018-0167-7.
To determine whether we could train convolutional neural network (CNN) models de novo with a small dataset, a total of 596 normal and abnormal ankle cases were collected and processed. Single- and multiview models were created to determine the effect of multiple views. Data augmentation was performed during training. The Inception V3, Resnet, and Xception convolutional neural networks were constructed utilizing the Python programming language with Tensorflow as the framework. Training was performed using single radiographic views. Measured output metrics were accuracy, positive predictive value (PPV), negative predictive value (NPV), sensitivity, and specificity. Model outputs were evaluated using both one and three radiographic views. Ensembles were created from a combination of CNNs after training. A voting method was implemented to consolidate the output from the three views and model ensemble. For single radiographic views, the ensemble of all 5 models produced the best accuracy at 76%. When all three views for a single case were utilized, the ensemble of all models resulted in the best output metrics with an accuracy of 81%. Despite our small dataset size, by utilizing an ensemble of models and 3 views for each case, we achieved an accuracy of 81%, which was in line with the accuracy of other models using a much higher number of cases with pre-trained models and models which implemented manual feature extraction.
为了确定我们是否可以使用小数据集从头开始训练卷积神经网络(CNN)模型,共收集和处理了 596 例正常和异常踝关节病例。创建了单视图和多视图模型以确定多视图的效果。在训练过程中进行了数据扩充。使用 Python 编程语言和 Tensorflow 作为框架构建了 Inception V3、Resnet 和 Xception 卷积神经网络。使用单张射线照片视图进行训练。测量的输出指标包括准确性、阳性预测值(PPV)、阴性预测值(NPV)、灵敏度和特异性。使用单张和三张射线照片视图评估了模型输出。在训练后,从 CNN 组合中创建了集成。实施投票方法来整合来自三个视图和模型集成的输出。对于单张射线照片视图,所有 5 个模型的集成产生了最佳的 76%准确率。当单个病例的所有三张视图都被使用时,所有模型的集成产生了最佳的输出指标,准确率为 81%。尽管我们的数据集规模较小,但通过使用模型集成和每个病例的 3 张视图,我们实现了 81%的准确率,与使用大量预训练模型和实现手动特征提取的模型的准确率相当。