“…Instruction-based learning has also been used in few-shot settings; popular variants include in-context learning, where the model's parameters are fixed and examples are provided as additional context (Brown et al, 2020;Lu et al, 2021;Kumar and Talukdar, 2021;Min et al, 2021), finetuning the entire model (Schick and Schütze, 2021a,c;Gao et al, 2021;Tam et al, 2021), and prompt tuning, where only the instruction itself is optimized (Shin et al, 2020;Hambardzumyan et al, 2021;Li and Liang, 2021;. Several works investigating the limitations and drawbacks of instruction-based few-shot approaches find that current LMs are mostly unable to understand complex instructions that go beyond short prompts or simple questions (Efrat and Levy, 2020;Weller et al, 2020;Webson and Pavlick, 2021) and that they are highly sensitive to the exact wording of the instructions provided (Jiang et al, 2020;Schick and Schütze, 2021a;Elazar et al, 2021). In a similar vein, Perez et al (2021) andLogan IV et al (2021) argue that prior work overestimates few-shot performance as manual prompt tuning is required to achieve good performance.…”