Walsh Reece, Osman Islam, Abdelaziz Omar, Shehata Mohamed S
Irving K. Barber Faculty of Science, University of British Columbia, Kelowna, BC V1V 1V7, Canada.
J Imaging. 2024 Jan 16;10(1):0. doi: 10.3390/jimaging10010023.
Few-shot learning aims to identify unseen classes with limited labelled data. Recent few-shot learning techniques have shown success in generalizing to unseen classes; however, the performance of these techniques has also been shown to degrade when tested on an out-of-domain setting. Previous work, additionally, has also demonstrated increasing reliance on supervised finetuning in an off-line or online capacity. This paper proposes a novel, fully self-supervised few-shot learning technique (FSS) that utilizes a vision transformer and masked autoencoder. The proposed technique can generalize to out-of-domain classes by finetuning the model in a fully self-supervised method for each episode. We evaluate the proposed technique using three datasets (all out-of-domain). As such, our results show that FSS has an accuracy gain of 1.05%, 0.12%, and 1.28% on the ISIC, EuroSat, and BCCD datasets, respectively, without the use of supervised training.
少样本学习旨在利用有限的标注数据识别未见类别。最近的少样本学习技术已在泛化到未见类别方面取得成功;然而,这些技术在域外设置上进行测试时,其性能也会下降。此外,先前的工作还表明对离线或在线能力的监督微调的依赖在增加。本文提出了一种新颖的、完全自监督的少样本学习技术(FSS),该技术利用视觉Transformer和掩码自动编码器。所提出的技术可以通过在每个情节中以完全自监督的方法微调模型来泛化到域外类别。我们使用三个数据集(均为域外数据集)对所提出的技术进行评估。因此,我们的结果表明,在不使用监督训练的情况下,FSS在ISIC、EuroSat和BCCD数据集上的准确率分别提高了1.05%、0.12%和1.28%。