“…For example, CLIP [81] adopts linear probing [12,31,32,109] and full-finetuning [25,31,48,99,101,109] when transferring to downstream tasks. Prompt adaptation of CLIP [63,81,105,112,114] is motivated by the success of prefix-tuning for language models [16,22,30,45,61,78,84,85,89]. Similarly, CLIP-Adapter [21] and Tip-Adapter [111] are inspired by parameter-efficient finetuning methods [39,44,110] that optimize lightweight MLPs while freezing the encoder.…”