Kuritcyn Petr, Kletzander Rosalie, Eisenberg Sophia, Wittenberg Thomas, Bruns Volker, Evert Katja, Keil Felix, Ziegler Paul K, Bankov Katrin, Wild Peter, Eckstein Markus, Hartmann Arndt, Geppert Carol I, Benz Michaela
Fraunhofer IIS, Fraunhofer Institute for Integrated Circuits IIS, Medical Image Analysis Group, Erlangen, Germany.
Institute of Pathology, University of Regensburg, Regensburg, Germany.
J Pathol Inform. 2024 Jun 6;15:100388. doi: 10.1016/j.jpi.2024.100388. eCollection 2024 Dec.
A vast multitude of tasks in histopathology could potentially benefit from the support of artificial intelligence (AI). Many examples have been shown in the literature and first commercial products with FDA or CE-IVDR clearance are available. However, two key challenges remain: (1) a scarcity of thoroughly annotated images, respectively the laboriousness of this task, and (2) the creation of robust models that can cope with the data heterogeneity in the field (domain generalization). In this work, we investigate how the combination of prototypical few-shot classification models and data augmentation can address both of these challenges. Based on annotated data sets that include multiple centers, multiple scanners, and two tumor entities, we examine the robustness and the adaptability of few-shot classifiers in multiple scenarios. We demonstrate that data from one scanner and one site are sufficient to train robust few-shot classification models by applying domain-specific data augmentation. The models achieved classification performance of around 90% on a multiscanner and multicenter database, which is on par with the accuracy achieved on the primary single-center single-scanner data. Various convolutional neural network (CNN) architectures can be used for feature extraction in the few-shot model. A comparison of nine state-of-the-art architectures yielded that EfficientNet B0 provides the best trade-off between accuracy and inference time. The classification of prototypical few-shot models directly relies on class prototypes derived from example images of each class. Therefore, we investigated the influence of prototypes originating from images from different scanners and evaluated their performance also on the multiscanner database. Again, our few-shot model showed a stable performance with an average absolute deviation in accuracy compared to the primary prototypes of 1.8% points. Finally, we examined the adaptability to a new tumor entity: classification of tissue sections containing urothelial carcinoma into normal, tumor, and necrotic regions. Only three annotations per subclass (e.g., muscle and adipose tissue are subclasses of normal tissue) were provided to adapt the few-shot model, which obtained an overall accuracy of 93.6%. These results demonstrate that prototypical few-shot classification is an ideal technology for realizing an interactive AI authoring system as it only requires few annotations and can be adapted to new tasks without involving retraining of the underlying feature extraction CNN, which would in turn require a selection of hyper-parameters based on data science expert knowledge. Similarly, it can be regarded as a guided annotation system. To this end, we realized a workflow and user interface that targets non-technical users.
组织病理学中的大量任务都可能从人工智能(AI)的支持中受益。文献中已经展示了许多例子,并且有首批获得FDA或CE-IVDR认证的商业产品。然而,仍然存在两个关键挑战:(1)缺乏经过充分注释的图像,以及这项任务的艰巨性;(2)创建能够应对该领域数据异质性的强大模型(领域泛化)。在这项工作中,我们研究了原型少样本分类模型和数据增强的结合如何应对这两个挑战。基于包含多个中心、多个扫描仪和两种肿瘤实体的注释数据集,我们在多种场景下检验了少样本分类器的稳健性和适应性。我们证明,通过应用特定领域的数据增强,来自一台扫描仪和一个站点的数据足以训练出强大的少样本分类模型。这些模型在多扫描仪和多中心数据库上实现了约90%的分类性能,这与在主要的单中心单扫描仪数据上所达到的准确率相当。在少样本模型中,可以使用各种卷积神经网络(CNN)架构进行特征提取。对九种先进架构的比较表明,EfficientNet B0在准确率和推理时间之间提供了最佳平衡。原型少样本模型的分类直接依赖于从每个类别的示例图像中导出的类原型。因此,我们研究了源自不同扫描仪图像的原型的影响,并在多扫描仪数据库上评估了它们的性能。同样,我们的少样本模型表现出稳定的性能,与主要原型相比,准确率的平均绝对偏差为1.8个百分点。最后,我们检验了对新肿瘤实体的适应性:将包含尿路上皮癌的组织切片分类为正常、肿瘤和坏死区域。每个子类仅提供三个注释(例如,肌肉和脂肪组织是正常组织的子类)来调整少样本模型,该模型获得了93.6%的总体准确率。这些结果表明,原型少样本分类是实现交互式AI创作系统的理想技术,因为它只需要少量注释,并且可以适应新任务,而无需对底层特征提取CNN进行重新训练,而重新训练又需要根据数据科学专家知识选择超参数。同样,它可以被视为一个引导式注释系统。为此,我们实现了一个面向非技术用户的工作流程和用户界面。