Singular Value Few-shot Adaptation of Vision-Language Models
Abstract
Vision-language models (VLMs) like CLIP have shown impressive zero-shot and few-shot learning capabilities across diverse applications. However, adapting these models to new fine-grained domains remains difficult due to reliance on prompt engineering and the high cost of full model fine-tuning. Existing adaptation approaches rely on augmented components, such as prompt tokens and adapter modules, which could limit adaptation quality, destabilize the model, and compromise the rich knowledge learned during pretraining. In this work, we present CLIP-SVD, a novel multi-modal and parameter-efficient adaptation technique that leverages Singular Value Decomposition (SVD) to modify the internal parameter space of CLIP without injecting additional modules. Specifically, we fine-tune only the singular values of the CLIP parameter matrices to rescale the basis vectors for domain adaptation while retaining the pretrained model. This design enables enhanced adaptation performance using only 0.04\% of the model's total parameters and better preservation of its generalization ability. CLIP-SVD achieves state-of-the-art classification results on 11 natural and 10 biomedical datasets, outperforming previous methods in both accuracy and generalization under few-shot settings. Additionally, we leverage a natural language-based approach to analyze the effectiveness and dynamics of the CLIP adaptation to allow interpretability of CLIP-SVD. The code is publicly available at https://github.com/HealthX-Lab/CLIP-SVD.
Community
CLIP-SVD is a parameter-efficient adaptation method for vision-language models that fine-tunes only the singular values of CLIP’s parameter matrices, enabling domain adaptation without added modules while preserving pretrained knowledge. It achieves state-of-the-art performance across natural and biomedical datasets in few-shot settings, with strong accuracy, generalization, and interpretability.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- GLAD: Generalizable Tuning for Vision-Language Models (2025)
- Hierarchical Cross-modal Prompt Learning for Vision-Language Models (2025)
- MSGCoOp: Multiple Semantic-Guided Context Optimization for Few-Shot Learning (2025)
- Attn-Adapter: Attention Is All You Need for Online Few-shot Learner of Vision-Language Model (2025)
- Accelerating Conditional Prompt Learning via Masked Image Modeling for Vision-Language Models (2025)
- One Last Attention for Your Vision-Language Model (2025)
- Towards Fine-Grained Adaptation of CLIP via a Self-Trained Alignment Score (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper