Meta-Adapter: An Online Few-shot Learner for Vision-Language Model Lin Song 2 Hang Wang 1
–Neural Information Processing Systems
The contrastive vision-language pre-training, known as CLIP, demonstrates remarkable potential in perceiving open-world visual concepts, enabling effective zero-shot image recognition. Nevertheless, few-shot learning methods based on CLIP typically require offline fine-tuning of the parameters on few-shot samples, resulting in longer inference time and the risk of over-fitting in certain domains. To tackle these challenges, we propose the Meta-Adapter, a lightweight residualstyle adapter, to refine the CLIP features guided by the few-shot samples in an online manner. With a few training samples, our method can enable effective few-shot learning capabilities and generalize to unseen data or tasks without additional fine-tuning, achieving competitive performance and high efficiency. Without bells and whistles, our approach outperforms the state-of-the-art online few-shot learning method by an average of 3.6% on eight image classification datasets with higher inference speed. Furthermore, our model is simple and flexible, serving as a plug-and-play module directly applicable to downstream tasks. Without further fine-tuning, Meta-Adapter obtains notable performance improvements in open-vocabulary object detection and segmentation tasks.
Neural Information Processing Systems
May-25-2025, 08:37:59 GMT
- Genre:
- Research Report (0.46)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning
- Neural Networks (0.46)
- Pattern Recognition (0.35)
- Statistical Learning (0.46)
- Natural Language > Large Language Model (0.37)
- Vision > Image Understanding (0.35)
- Machine Learning
- Information Technology > Artificial Intelligence