“…In contrast, recent works have focused on more quantitative criteria: correlation between explainability methods for measuring consistency Wallace, 2019, Serrano andSmith, 2019], sufficiency and comprehensiveness [DeYoung et al, 2020], and simulability: whether a human or machine consumer of explanations understands the model behavior well enough to predict its output on unseen examples [Doshi-Velez and Kim, 2017]. Simulability, in particular, has a number of desirable properties, such as being intuitively aligned with the goal of communicating the underlying model behavior to humans and being measurable in manual and automated experiments [Treviso and Martins, 2020, Hase and Bansal, 2020, Pruthi et al, 2020. Figure 1: Illustration of our SMaT framework.…”