Faculty of Applied Medical Sciences, Jordan University of Science and Technology, Irbid 22110, Jordan.
Medical Image Optimisation and Perception Group (MIOPeG), Faculty of Health Sciences, The University of Sydney, Sydney, New South Wales, Australia.
Comput Biol Med. 2018 May 1;96:294-298. doi: 10.1016/j.compbiomed.2018.04.005. Epub 2018 Apr 13.
While there is much literature describing the radiologic detection of breast cancer, there are limited data available on the agreement between experts when delineating and classifying breast lesions. The aim of this work is to measure the level of agreement between expert radiologists when delineating and classifying breast lesions as demonstrated through Breast Imaging Reporting and Data System (BI-RADS) and quantitative shape metrics.
Forty mammographic images, each containing a single lesion, were presented to nine expert breast radiologists using a high specification interactive digital drawing tablet with stylus. Each reader was asked to manually delineate the breast masses using the tablet and stylus and then visually classify the lesion according to the American College of Radiology (ACR) BI-RADS lexicon. The delineated lesion compactness and elongation were computed using Matlab software. Intraclass Correlation Coefficient (ICC) and Cohen's kappa were used to assess inter-observer agreement for delineation and classification outcomes, respectively.
Inter-observer agreement was fair for BI-RADS shape (kappa = 0.37) and moderate for margin (kappa = 0.58) assessments. Agreement for quantitative shape metrics was good for lesion elongation (ICC = 0.82) and excellent for compactness (ICC = 0.93).
Fair to moderate levels of agreement was shown by radiologists for shape and margin classifications of cancers using the BI-RADS lexicon. When quantitative shape metrics were used to evaluate radiologists' delineation of lesions, good to excellent inter-observer agreement was found. The results suggest that qualitative descriptors such as BI-RADS lesion shape and margin understate the actual level of expert radiologist agreement.
虽然有很多文献描述了乳腺癌的放射学检测,但关于专家在描绘和分类乳腺病变时的一致性的数据有限。本研究旨在通过乳腺影像报告和数据系统(BI-RADS)和定量形状指标来衡量专家放射科医生在描绘和分类乳腺病变时的一致性水平。
使用具有触笔的高规格交互式数字绘图板向 9 位专家乳腺放射科医生展示了 40 张乳腺图像,每张图像均包含单个病变。要求每位读者使用平板电脑和触笔手动描绘乳腺肿块,然后根据美国放射学院(ACR)BI-RADS 词汇表对病变进行视觉分类。使用 Matlab 软件计算描绘的病变紧凑度和伸长率。使用组内相关系数(ICC)和 Cohen's kappa 分别评估描绘和分类结果的观察者间一致性。
BI-RADS 形状(kappa=0.37)和边缘(kappa=0.58)评估的观察者间一致性为中等,而定量形状指标的一致性为良好(ICC=0.82)和优秀(ICC=0.93)。
放射科医生使用 BI-RADS 词汇对癌症的形状和边缘分类表现出中等至中等偏上的一致性。当使用定量形状指标评估放射科医生对病变的描绘时,发现观察者间具有良好至极好的一致性。结果表明,BI-RADS 病变形状和边缘等定性描述低估了专家放射科医生的实际一致性水平。