Gao Yue, Zeng Shaoqing, Xu Xiaoyan, Li Huayi, Yao Shuzhong, Song Kun, Li Xiao, Chen Lingxi, Tang Junying, Xing Hui, Yu Zhiying, Zhang Qinghua, Zeng Shue, Yi Cunjian, Xie Hongning, Xiong Xiaoming, Cai Guangyao, Wang Zhi, Wu Yuan, Chi Jianhua, Jiao Xiaofei, Qin Yan, Mao Xiaogang, Chen Yu, Jin Xin, Mo Qingqing, Chen Pingbo, Huang Yi, Shi Yushuang, Wang Junmei, Zhou Yimin, Ding Shuping, Zhu Shan, Liu Xin, Dong Xiangyi, Cheng Lin, Zhu Linlin, Cheng Huanhuan, Cha Li, Hao Yanli, Jin Chunchun, Zhang Ludan, Zhou Peng, Sun Meng, Xu Qin, Chen Kehua, Gao Zeyan, Zhang Xu, Ma Yuanyuan, Liu Yan, Xiao Liling, Xu Li, Peng Lin, Hao Zheyu, Yang Mi, Wang Yane, Ou Hongping, Jia Yongmei, Tian Lihua, Zhang Wei, Jin Ping, Tian Xun, Huang Lei, Wang Zhen, Liu Jiahao, Fang Tian, Yan Danmei, Cao Heng, Ma Jingjing, Li Xiaoting, Zheng Xu, Lou Hua, Song Chunyan, Li Ruyuan, Wang Siyuan, Li Wenqian, Zheng Xulei, Chen Jing, Li Guannan, Chen Ruqi, Xu Cheng, Yu Ruidi, Wang Ji, Xu Sen, Kong Beihua, Xie Xing, Ma Ding, Gao Qinglei
National Clinical Research Center for Obstetrics and Gynaecology, Cancer Biology Research Centre (Key Laboratory of the Ministry of Education) and Department of Gynaecology and Obstetrics, Tongji Hospital, Wuhan, China.
Department of Obstetrics and Gynecology, First Affiliated Hospital of Sun Yat-sen University, Guangzhou, China.
Lancet Digit Health. 2022 Mar;4(3):e179-e187. doi: 10.1016/S2589-7500(21)00278-8.
Ultrasound is a critical non-invasive test for preoperative diagnosis of ovarian cancer. Deep learning is making advances in image-recognition tasks; therefore, we aimed to develop a deep convolutional neural network (DCNN) model that automates evaluation of ultrasound images and to facilitate a more accurate diagnosis of ovarian cancer than existing methods.
In this retrospective, multicentre, diagnostic study, we collected pelvic ultrasound images from ten hospitals across China between September 2003, and May 2019. We included consecutive adult patients (aged ≥18 years) with adnexal lesions in ultrasonography and healthy controls and excluded duplicated cases and patients without adnexa or pathological diagnosis. For DCNN model development, patients were assigned to the training dataset (34 488 images of 3755 patients with ovarian cancer, 541 442 images of 101 777 controls). For model validation, patients were assigned to the internal validation dataset (3031 images of 266 patients with ovarian cancer, 5385 images of 602 with benign adnexal lesions), external validation datasets 1 (486 images of 67 with ovarian cancer, 933 images of 268 with benign adnexal lesions), and 2 (1253 images of 166 with ovarian cancer, 5257 images of 723 benign adnexal lesions). Using these datasets, we assessed the diagnostic value of DCNN, compared DCNN with 35 radiologists, and explored whether DCNN could augment the diagnostic accuracy of six radiologists. Pathological diagnosis was the reference standard.
For DCNN to detect ovarian cancer, AUC was 0·911 (95% CI 0·886-0·936) in the internal dataset, 0·870 (95% CI 0·822-0·918) in external validation dataset 1, and 0·831 (95% CI 0·793-0·869) in external validation dataset 2. The DCNN model was more accurate than radiologists at detecting ovarian cancer in the internal dataset (88·8% vs 85·7%) and external validation dataset 1 (86·9% vs 81·1%). Accuracy and sensitivity of diagnosis increased more after DCNN-assisted diagnosis than assessment by radiologists alone (87·6% [85·0-90·2] vs 78·3% [72·1-84·5], p<0·0001; 82·7% [78·5-86·9] vs 70·4% [59·1-81·7], p<0·0001). The average accuracy of DCNN-assisted evaluations for six radiologists reached 0·876 and were significantly augmented when they were DCNN-assisted (p<0·05).
The performance of DCNN-enabled ultrasound exceeded the average diagnostic level of radiologists matched the level of expert ultrasound image readers, and augmented radiologists' accuracy. However, these observations warrant further investigations in prospective studies or randomised clinical trials.
National Key Basic Research Program of China, National Sci-Tech Support Projects, and National Natural Science Foundation of China.
超声是卵巢癌术前诊断的关键非侵入性检查。深度学习在图像识别任务中取得了进展;因此,我们旨在开发一种深度卷积神经网络(DCNN)模型,以实现超声图像评估的自动化,并比现有方法更准确地诊断卵巢癌。
在这项回顾性、多中心诊断研究中,我们收集了2003年9月至2019年5月期间中国十家医院的盆腔超声图像。我们纳入了超声检查中有附件病变的连续成年患者(年龄≥18岁)和健康对照,并排除了重复病例以及没有附件或病理诊断的患者。对于DCNN模型开发,患者被分配到训练数据集(3755例卵巢癌患者的34488张图像,101777例对照的541442张图像)。对于模型验证,患者被分配到内部验证数据集(266例卵巢癌患者的3031张图像,602例良性附件病变患者的5385张图像)、外部验证数据集1(67例卵巢癌患者的486张图像,268例良性附件病变患者的933张图像)和外部验证数据集2(166例卵巢癌患者的1253张图像,723例良性附件病变患者的5257张图像)。使用这些数据集,我们评估了DCNN的诊断价值,将DCNN与35名放射科医生进行了比较,并探讨了DCNN是否可以提高6名放射科医生的诊断准确性。病理诊断为参考标准。
对于DCNN检测卵巢癌,内部数据集的AUC为0.911(95%CI 0.886 - 0.936),外部验证数据集1为0.870(95%CI 0.822 - 0.918),外部验证数据集2为0.831(95%CI 0.793 - 0.869)。DCNN模型在内部数据集(88.8%对85.