“…Combining autoregressive generation with transformers (Devlin et al, 2018; has been of enormous impact in language modelling Rae et al, 2021), protein folding (Jumper et al, 2021), vision-language models (Alayrac et al, 2022;Tsimpoukelli et al, 2021;, code generation (Chen et al, 2021c;Li et al, 2022b), dialogue systems with retrieval capabilities (Nakano et al, 2021;Thoppilan et al, 2022), speech recognition (Pratap et al, 2020), neural machine translation (Johnson et al, 2019) and more (Bommasani et al, 2021). Recently researchers have explored task decomposition and grounding with language models (Ahn et al, 2022;Huang et al, 2022). Li et al (2022a) construct a control architecture, consisting of a sequence tokenizer, a pretrained language model and a task-specific feed-forward network.…”