Few-shot learning aims to learn a classifier with limited training instances to recognize unseen classes in test. Recently, some effective few-shot learning approaches have achieved promising classification performance. However, these approaches do not pay attention to the importance of task-relevant features in image classification. To address this issue, we propose Task-Relevant Graph Metric Learning method that adopts a novel meta-learning framework for transductive inference. Specifically, we first extract task-relevant features by a Squeeze-and-Excitation module. Then learning a graph construction module in order to obtain the manifold structure in the data. Afterward, self-training is utilized to propagate labels from labeled instances to unlabeled test instances. Experiments on benchmark datasets demonstrate that TRGML improves classification performance (4%-5%) over baseline systems on miniImageNet and tieredImageNet.
Semantic segmentation with deep learning has achieved remarkable progress in classifying the pixels in the image. Acquiring sufficient ground-truth supervision to train deep visual models has been a bottleneck over the years due to the data-hungry nature of deep learning. Image-level label-based weakly supervised semantic segmentation (WSSS) aims to adopt image-level labels to train semantic segmentation models, saving vast human labors for costly pixel-level annotations. A typical pipeline for this problem is to adopt Class Activation Maps (CAMs) with image-level labels to generate pseudo-masks (a.k.a. seeds) and then use them for training segmentation models. The main difficulty is that seeds are usually sparse and incomplete. In recent years, GCNs have made great strides in various fields. GCN can perform global modeling and reasoning on the relationship between regions, which is beneficial for many computer vision tasks. This is our motivation to combine these two aspects. Therefore we propose the module of class-related graph convolution. Because there are differences between classes, our GCN is parallel. Each GCN can learn the classrelated region extension strategy. To enable GCN to learn more authentic relationships, we also introduce the attention mechanisms. We conduct lots of experiments on the public PASCAL VOC dataset, and our model yields state-of-the-art performance.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.