TY - JOUR
T1 - Feature hallucination via Maximum A Posteriori for few-shot learning[Formula presented]
AU - Wu, Jiaying
AU - Dong, Ning
AU - Liu, Fan
AU - Yang, Sai
AU - Hu, Jinglu
N1 - Funding Information:
This work is partially supported by China Scholarship Council (No. 201908320571 ), Natural Science Foundation of Jiangsu Province under Grant No. BK20191298 , Fundamental Research Funds for the Central Universities under Grant No. B200202175 , and Key Laboratory of Coastal Disaster and Protection of Ministry of Euducation, Hohai University , under Grant No. 201905 .
Publisher Copyright:
© 2021 Elsevier B.V.
PY - 2021/8/5
Y1 - 2021/8/5
N2 - Few-shot learning aims to train an effective classifier in a small data regime. Due to the scarcity of training samples (usually as small as 1 or 5), traditional deep learning solutions often suffer from overfitting. To address this issue, an intuitive idea is to augment or hallucinate sufficient training data. For this purpose, in this paper, we propose a simple yet effective method to build a model for novel categories with few samples. Specifically, we assume that each category in the base set follows a Gaussian distribution, so that we can employ Maximum A Posteriori (MAP) to estimate the distribution of a novel category with even one example. To achieve this goal, we first transform each base category into Gaussian form with power transformation for MAP estimation. Then, we estimate the Gaussian mean of the novel category under the Gaussian prior given few samples from it. Finally, each novel category is represented by a unique Gaussian distribution, where sufficient trainable features can be sampled to obtain a highly accurate classifier for final predictions. Experimental results on four few-shot benchmarks show that it significantly outperforms the baseline methods on both 1- and 5-shot tasks. Extensive results on cross-domain tasks and visualization of estimated feature distribution also demonstrate its effectiveness.
AB - Few-shot learning aims to train an effective classifier in a small data regime. Due to the scarcity of training samples (usually as small as 1 or 5), traditional deep learning solutions often suffer from overfitting. To address this issue, an intuitive idea is to augment or hallucinate sufficient training data. For this purpose, in this paper, we propose a simple yet effective method to build a model for novel categories with few samples. Specifically, we assume that each category in the base set follows a Gaussian distribution, so that we can employ Maximum A Posteriori (MAP) to estimate the distribution of a novel category with even one example. To achieve this goal, we first transform each base category into Gaussian form with power transformation for MAP estimation. Then, we estimate the Gaussian mean of the novel category under the Gaussian prior given few samples from it. Finally, each novel category is represented by a unique Gaussian distribution, where sufficient trainable features can be sampled to obtain a highly accurate classifier for final predictions. Experimental results on four few-shot benchmarks show that it significantly outperforms the baseline methods on both 1- and 5-shot tasks. Extensive results on cross-domain tasks and visualization of estimated feature distribution also demonstrate its effectiveness.
KW - Feature hallucination
KW - Few-shot learning
KW - Gaussian prior
KW - Maximum A Posteriori
UR - http://www.scopus.com/inward/record.url?scp=85105568443&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85105568443&partnerID=8YFLogxK
U2 - 10.1016/j.knosys.2021.107129
DO - 10.1016/j.knosys.2021.107129
M3 - Article
AN - SCOPUS:85105568443
SN - 0950-7051
VL - 225
JO - Knowledge-Based Systems
JF - Knowledge-Based Systems
M1 - 107129
ER -