Togo Ren, Watanabe Haruna, Ogawa Takahiro, Haseyama Miki
Education and Research Center for Mathematical and Data Science, Hokkaido University, N-12, W-7, Kita-ku, Sapporo, 060-0812, Japan.
Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
Comput Biol Med. 2020 Aug;123:103903. doi: 10.1016/j.compbiomed.2020.103903. Epub 2020 Jul 8.
The aim of this study was to determine whether our deep convolutional neural network-based anomaly detection model can distinguish differences in esophagus images and stomach images obtained from gastric X-ray examinations.
A total of 6012 subjects were analyzed as our study subjects. Since the number of esophagus X-ray images is much smaller than the number of gastric X-ray images taken in X-ray examinations, we took an anomaly detection approach to realize the task of organ classification. We constructed a deep autoencoding gaussian mixture model (DAGMM) with a convolutional autoencoder architecture. The trained model can produce an anomaly score for a given test X-ray image. For comparison, the original DAGMM, AnoGAN, and a One-Class Support Vector Machine (OCSVM) that were trained with features obtained by a pre-trained Inception-v3 network were used.
Sensitivity, specificity, and the calculated harmonic mean of the proposed method were 0.956, 0.980, and 0.968, respectively. Those of the original DAGMM were 0.932, 0.883, and 0.907, respectively. Those of AnoGAN were 0.835, 0.833, and 0.834, respectively, and those of OCSVM were 0.932, 0.935, and 0.934, respectively. Experimental results showed the effectiveness of the proposed method for an organ classification task.
Our deep convolutional neural network-based anomaly detection model has shown the potential for clinical use in organ classification.
本研究旨在确定基于深度卷积神经网络的异常检测模型能否区分胃X线检查中获得的食管图像和胃图像之间的差异。
共分析6012名受试者作为研究对象。由于食管X线图像数量远少于X线检查中拍摄的胃X线图像数量,我们采用异常检测方法来实现器官分类任务。我们构建了一个具有卷积自动编码器架构的深度自动编码高斯混合模型(DAGMM)。训练后的模型可以为给定的测试X线图像生成异常分数。为作比较,使用了原始的DAGMM、AnoGAN以及一个使用预训练的Inception-v3网络获得的特征进行训练的单类支持向量机(OCSVM)。
所提方法的灵敏度、特异度及计算得到的调和均值分别为0.956、0.980和0.968。原始DAGMM的分别为0.932、0.883和0.907。AnoGAN的分别为0.835、0.833和0.834,OCSVM的分别为0.932、0.935和0.934。实验结果表明所提方法在器官分类任务中的有效性。
我们基于深度卷积神经网络的异常检测模型已显示出在器官分类临床应用中的潜力。