Department of Computer Science and Engineering, United International University, Dhaka, 1212, Bangladesh.
Department of Computer Science and Engineering, Bangladesh University of Business and Technology, Mirpur, Dhaka, 1216, Bangladesh.
Sci Rep. 2024 Nov 9;14(1):27358. doi: 10.1038/s41598-024-78363-w.
Breast cancer is the second most common type of cancer among women. Prompt detection of breast cancer can impede its advancement to more advanced phases, thereby elevating the probability of favorable treatment consequences. Histopathological images are commonly used for breast cancer classification due to their detailed cellular information. Existing diagnostic approaches rely on Convolutional Neural Networks (CNNs) which are limited to local context resulting in a lower classification accuracy. Therefore, we present a fusion model composed of a Vision Transformer (ViT) and custom Atrous Spatial Pyramid Pooling (ASPP) network with an attention mechanism for effectively classifying breast cancer from histopathological images. ViT enables the model to attain global features, while the ASPP network accommodates multiscale features. Fusing the features derived from the models resulted in a robust breast cancer classifier. With the help of five-stage image preprocessing technique, the proposed model achieved 100% accuracy in classifying breast cancer on the BreakHis dataset at 100X and 400X magnification factors. On 40X and 200X magnifications, the model achieved 99.25% and 98.26% classification accuracy respectively. With a commendable classification efficacy on histopathological images, the model can be considered a dependable option for proficient breast cancer classification.
乳腺癌是女性中第二常见的癌症类型。乳腺癌的早期发现可以阻止其向更晚期发展,从而提高治疗效果的可能性。由于组织病理学图像具有详细的细胞信息,因此通常用于乳腺癌分类。现有的诊断方法依赖于卷积神经网络 (CNN),但由于 CNN 仅限于局部上下文,因此分类准确性较低。因此,我们提出了一种由 Vision Transformer (ViT) 和自定义的 Atrous Spatial Pyramid Pooling (ASPP) 网络组成的融合模型,该模型具有注意力机制,可有效从组织病理学图像中分类乳腺癌。ViT 使模型能够获得全局特征,而 ASPP 网络则适应多尺度特征。融合来自模型的特征导致了一个强大的乳腺癌分类器。在五阶段图像预处理技术的帮助下,所提出的模型在 BreakHis 数据集上在 100X 和 400X 放大倍数下实现了 100%的乳腺癌分类准确率。在 40X 和 200X 放大倍数下,模型分别实现了 99.25%和 98.26%的分类准确率。该模型在组织病理学图像上具有出色的分类效果,可以被认为是一种用于熟练分类乳腺癌的可靠选择。