Papers
arxiv:2307.04114

FILM: How can Few-Shot Image Classification Benefit from Pre-Trained Language Models?

Published on Jul 9, 2023
Authors:
,
,
,

Abstract

A novel few-shot learning framework uses pre-trained language models with contrastive learning, a metric module for cosine similarity, and MAML for bi-level optimization to improve generalization and transferability.

AI-generated summary

Few-shot learning aims to train models that can be generalized to novel classes with only a few samples. Recently, a line of works are proposed to enhance few-shot learning with accessible semantic information from class names. However, these works focus on improving existing modules such as visual prototypes and feature extractors of the standard few-shot learning framework. This limits the full potential use of semantic information. In this paper, we propose a novel few-shot learning framework that uses pre-trained language models based on contrastive learning. To address the challenge of alignment between visual features and textual embeddings obtained from text-based pre-trained language model, we carefully design the textual branch of our framework and introduce a metric module to generalize the cosine similarity. For better transferability, we let the metric module adapt to different few-shot tasks and adopt MAML to train the model via bi-level optimization. Moreover, we conduct extensive experiments on multiple benchmarks to demonstrate the effectiveness of our method.

Community

Sign up or log in to comment

Get this paper in your agent:

hf papers read 2307.04114
Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2307.04114 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2307.04114 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2307.04114 in a Space README.md to link it from this page.

Collections including this paper 1