Interpretable Few-Shot Learning with Contrastive Constraint

Publication Type:
Journal Article
Citation:
Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58, (12), pp. 2573-2584
Issue Date:
2021-12-01
Filename Description Size
基于对比约束的可解释小样本学习.pdfPublished version3.2 MB
Adobe PDF
Full metadata record
Different from deep learning with large scale supervision, few-shot learning aims to learn the samples' characteristics from a few labeled examples. Apparently, few-shot learning is more in line with the visual cognitive mechanism of the human brain. In recent years, few-shot learning has attracted more researchers' attention. In order to discover the semantic similarities between the query set (unlabeled image) and support set (few labeled images) in feature embedding space, methods which combine meta-learning and metric learning have emerged and achieved great performance on few-shot image classification tasks. However, these methods lack the interpretability, which means they could not provide a reasoning explainable process like human cognitive mechanism. Therefore, we propose a novel interpretable few-shot learning method called INT-FSL based on the positional attention mechanism, which aims to reveal two key problems in few-shot classification: 1)Which parts of the unlabeled image play an important role in classification task; 2)Which class of features reflected by the key parts. Besides, we design the contrastive constraints on global and local levels in every few-shot meta task, for alleviating the limited supervision with the internal information of the data. We conduct extensive experiments on three image benchmark datasets. The results show that the proposed model INT-FSL not only could improve the classification performance on few-shot learning effectively, but also has good interpretability in the reasoning process.
Please use this identifier to cite or link to this item: