Abstract: Vision-Language models (VLMs) like Contrastive Language-Image Pre-Training (CLIP) have been extensively adapted for few-shot classification. Most few-shot methods rely on randomly selected ...
Some results have been hidden because they may be inaccessible to you
Show inaccessible results