“…Previous efforts to train models to use explanations (Mishra et al, 2022 ), whether from scratch (Camburu et al, 2018 , Lampinen, Roy, et al, 2022 ), through fine-tuning (Lampinen, Dasgupta, et al, 2022 ), or through conditioning with in-context prompts at evaluation time (Lu et al, 2022 ; Wei et al, 2022 ), have shown improved performance over models without explicit explanations. However, much of the existing literature remains largely empirical with limited theoretic accounts for the phenomenon (Xie et al, 2021 ).…”