“…More specifically, (Li et al, 2019) infers a nonlinear interaction cost function from observed noisy matching data via optimal transport while (Trivedi et al, 2020) learns a reward function which serves as an explanation for an observed snapshot of a graph via inverse reinforcement learning. Alternatively, other works including (Ling et al, 2018;Li et al, 2020;Heaton et al, 2021) present a learning-theoretic approach, where an operator approximating the game gradient is estimated within a deep learning framework, to infer the unknown utility parameters from equilibria observations. (Ling et al, 2018) and (Ling et al, 2019) propose a parametric end-to-end framework to learn the utilities in a two-player zero-sum game from an observed quantal response equilibrium and nested logit quan-tal response equilibrium, respectively, using a differentiable game solver layer.…”