- Library Home /
- Search Collections /
- Open Collections /
- Browse Collections /
- UBC Faculty Research and Publications /
- Fully Self-Supervised Out-of-Domain Few-Shot Learning...
Open Collections
UBC Faculty Research and Publications
Fully Self-Supervised Out-of-Domain Few-Shot Learning with Masked Autoencoders Walsh, Reece; Osman, Islam; Abdelaziz, Omar; Shehata, Mohamed S.
Abstract
Few-shot learning aims to identify unseen classes with limited labelled data. Recent few-shot learning techniques have shown success in generalizing to unseen classes; however, the performance of these techniques has also been shown to degrade when tested on an out-of-domain setting. Previous work, additionally, has also demonstrated increasing reliance on supervised finetuning in an off-line or online capacity. This paper proposes a novel, fully self-supervised few-shot learning technique (FSS) that utilizes a vision transformer and masked autoencoder. The proposed technique can generalize to out-of-domain classes by finetuning the model in a fully self-supervised method for each episode. We evaluate the proposed technique using three datasets (all out-of-domain). As such, our results show that FSS has an accuracy gain of 1.05%, 0.12%, and 1.28% on the ISIC, EuroSat, and BCCD datasets, respectively, without the use of supervised training.
Item Metadata
Title |
Fully Self-Supervised Out-of-Domain Few-Shot Learning with Masked Autoencoders
|
Creator | |
Publisher |
Multidisciplinary Digital Publishing Institute
|
Date Issued |
2024-01-16
|
Description |
Few-shot learning aims to identify unseen classes with limited labelled data. Recent few-shot learning techniques have shown success in generalizing to unseen classes; however, the performance of these techniques has also been shown to degrade when tested on an out-of-domain setting. Previous work, additionally, has also demonstrated increasing reliance on supervised finetuning in an off-line or online capacity. This paper proposes a novel, fully self-supervised few-shot learning technique (FSS) that utilizes a vision transformer and masked autoencoder. The proposed technique can generalize to out-of-domain classes by finetuning the model in a fully self-supervised method for each episode. We evaluate the proposed technique using three datasets (all out-of-domain). As such, our results show that FSS has an accuracy gain of 1.05%, 0.12%, and 1.28% on the ISIC, EuroSat, and BCCD datasets, respectively, without the use of supervised training.
|
Subject | |
Genre | |
Type | |
Language |
eng
|
Date Available |
2024-03-04
|
Provider |
Vancouver : University of British Columbia Library
|
Rights |
CC BY 4.0
|
DOI |
10.14288/1.0440608
|
URI | |
Affiliation | |
Citation |
Journal of Imaging 10 (1): 23 (2024)
|
Publisher DOI |
10.3390/jimaging10010023
|
Peer Review Status |
Reviewed
|
Scholarly Level |
Faculty
|
Rights URI | |
Aggregated Source Repository |
DSpace
|
Item Media
Item Citations and Data
Rights
CC BY 4.0