“…• They are flexible in terms of the target distribution: Stein discrepancy requires the knowledge of the pdfs up to normalizing constant only, the Wasserstein distance can be written explicitly as a function of the inverse cdf (see (2.7)), the mean embedding (which is the underlying representation of probability measures used in MMD) can be computed analytically for different target distribution and kernel pairs. • These discrepancy measures show excellent performance in various (complementary) applications such as model criticism [37,31], two-sample [24,20], independence [21,46] and goodness-of-fit testing [36,8,28,17,26,19,15,1], portfolio valuation [5], statistical inference of generative models [6] and post selection inference [59], causal discovery [41,46], generative adversarial networks [11,35,3], assessing and tuning MCMC samplers [16,17,26,18], or designing Monte-Carlo control functionals for variance reduction [44,43,52], among many others.…”