Stanitsas Panagiotis, Cherian Anoop, Morellas Vassilios, Tejpaul Resha, Papanikolopoulos Nikolaos, Truskinovsky Alexander
Department of Computer Science and Engineering, University of Minnesota, Minneapolis, MN, United States.
Australian Center for Robotic Vision, Australian National University, Canberra, ACT, Australia.
Front Digit Health. 2020 Dec;2. doi: 10.3389/fdgth.2020.572671. Epub 2020 Dec 7.
Cancerous Tissue Recognition (CTR) methodologies are continuously integrating advancements at the forefront of machine learning and computer vision, providing a variety of inference schemes for histopathological data. Histopathological data, in most cases, come in the form of high-resolution images, and thus methodologies operating at the patch level are more computationally attractive. Such methodologies capitalize on pixel level annotations (tissue delineations) from expert pathologists, which are then used to derive labels at the patch level. In this work, we envision a digital connected health system that augments the capabilities of the clinicians by providing powerful feature descriptors that may describe malignant regions.
We start with a patch level descriptor, termed Covariance-Kernel Descriptor (CKD), capable of compactly describing tissue architectures associated with carcinomas. To leverage the recognition capability of the CKDs to larger slide regions, we resort to a multiple instance learning framework. In that direction, we derive the Weakly Annotated Image Descriptor (WAID) as the parameters of classifier decision boundaries in a Multiple Instance Learning framework. The WAID is computed on bags of patches corresponding to larger image regions for which binary labels (malignant vs. benign) are provided, thus obviating the necessity for tissue delineations.
The CKD was seen to outperform all the considered descriptors, reaching classification accuracy (ACC) of 92.83%. and area under the curve (AUC) of 0.98. The CKD captures higher order correlations between features and was shown to achieve superior performance against a large collection of computer vision features on a private breast cancer dataset. The WAID outperform all other descriptors on the Breast Cancer Histopathological database (BreakHis) where correctly classified malignant (CCM) instances reached 91.27 and 92.00% at the patient and image level, respectively, without resorting to a deep learning scheme achieves state-of-the-art performance.
Our proposed derivation of the CKD and WAID can help medical experts accomplish their work accurately and faster than the current state-of-the-art.
癌组织识别(CTR)方法不断融合机器学习和计算机视觉前沿的进展,为组织病理学数据提供了多种推理方案。在大多数情况下,组织病理学数据以高分辨率图像的形式呈现,因此在补丁级别运行的方法在计算上更具吸引力。此类方法利用专家病理学家的像素级注释(组织轮廓),然后用于在补丁级别得出标签。在这项工作中,我们设想了一个数字互联健康系统,通过提供可描述恶性区域的强大特征描述符来增强临床医生的能力。
我们从一个补丁级别的描述符开始,称为协方差核描述符(CKD),它能够紧凑地描述与癌相关的组织结构。为了将CKD的识别能力扩展到更大的玻片区域,我们采用了多实例学习框架。在这个方向上,我们将弱注释图像描述符(WAID)推导为多实例学习框架中分类器决策边界的参数。WAID是在对应于较大图像区域的补丁包上计算的,这些区域提供了二元标签(恶性与良性),从而无需组织轮廓。
CKD被证明优于所有考虑的描述符,分类准确率(ACC)达到92.83%,曲线下面积(AUC)为0.98。CKD捕获了特征之间的高阶相关性,并在一个私人乳腺癌数据集上针对大量计算机视觉特征显示出卓越的性能。在乳腺癌组织病理学数据库(BreakHis)上,WAID优于所有其他描述符,在不采用深度学习方案的情况下,在患者和图像级别正确分类的恶性(CCM)实例分别达到91.27%和92.00%,实现了当前最优性能。
我们提出的CKD和WAID的推导可以帮助医学专家比当前的最优方法更准确、更快地完成工作。