Department of Diagnostic Imaging, Rhode Island Hospital, Alpert Medical School of Brown University, Providence, RI.
Mallinckrodt Institute of Radiology, School of Medicine, Washington University, St. Louis, MO.
Ultrasound Q. 2020 Jun;36(2):164-172. doi: 10.1097/RUQ.0000000000000501.
This study evaluates the performance of convolutional neural networks (CNNs) in risk stratifying the malignant potential of thyroid nodules alongside traditional methods such as American College of Radiology Thyroid Imaging Reporting and Data System (ACR TIRADS). The data set consisted of 651 pathology-proven thyroid nodules (500 benign, 151 malignant) from 571 patients collected at a single tertiary academic medical center. Each thyroid nodule consisted of two orthogonal views (sagittal and transverse) for a total of 1,302 grayscale images. A CNN classifier was developed to identify malignancy versus benign thyroid nodules, and a nested double cross validation scheme was applied to allow for both model parameter selection and for model accuracy evaluation. All thyroid nodules were classified according to ACR TIRADS criteria and were compared with their respective CNN-generated malignancy scores. The best performing model was the MobileNet CNN ensemble with an area under the curve of 0.86 (95% confidence interval, 0.83-0.90). Thyroid nodules within the highest and lowest CNN risk strata had malignancy rates of 81.4% and 5.9%, respectively. The rate of malignancy for ACR TIRADS ranged from 0% for TR1 nodules to 60% for TR5 nodules. Convolutional neural network malignancy scores correlated well with TIRADS levels, as malignancy scores ranged from 0.194 for TR1 nodules and 0.519 for TR5 nodules. Convolutional neural networks can be trained to generate accurate malignancy risk scores for thyroid nodules. These predictive models can aid in risk stratifying thyroid nodules alongside traditional professional guidelines such as TIRADS and can function as an adjunct tool for the radiologist when identifying those patients requiring further histopathologic workup.
本研究评估了卷积神经网络(CNN)在与美国放射学院甲状腺影像报告和数据系统(ACR TIRADS)等传统方法一起对甲状腺结节恶性潜能进行风险分层方面的性能。该数据集由来自单一三级学术医疗中心的 571 名患者的 651 个经病理证实的甲状腺结节(500 个良性,151 个恶性)组成。每个甲状腺结节包含两个正交视图(矢状面和横断面),总共有 1302 个灰度图像。开发了一个 CNN 分类器来识别恶性与良性甲状腺结节,并应用嵌套的双交叉验证方案来允许模型参数选择和模型准确性评估。所有甲状腺结节均根据 ACR TIRADS 标准进行分类,并与各自的 CNN 生成的恶性评分进行比较。表现最佳的模型是 MobileNet CNN 集成,曲线下面积为 0.86(95%置信区间,0.83-0.90)。最高和最低 CNN 风险分层的甲状腺结节恶性率分别为 81.4%和 5.9%。ACR TIRADS 的恶性率从 TR1 结节的 0%到 TR5 结节的 60%不等。卷积神经网络的恶性评分与 TIRADS 水平相关性良好,恶性评分范围从 TR1 结节的 0.194 到 TR5 结节的 0.519。卷积神经网络可以被训练来为甲状腺结节生成准确的恶性风险评分。这些预测模型可以与传统的专业指南(如 TIRADS)一起帮助对甲状腺结节进行风险分层,并可以作为放射科医生在确定需要进一步组织病理学检查的患者时的辅助工具。