“…An open question in AI research and development is how to represent and specify ethical choices and constraints for this class of technologies in computational terms [Ajmeri et al, 2020;Amodei et al, 2016;Awad et al, 2022;Dignum, 2017; Version with Appendix: https://arxiv.org/abs/2301.08491 Code: https://github.com/Liza-Tennant/moral choice dyadic Yu et al, 2018;Wallach, 2010]. In particular, there is an increasing interest in understanding how certain types of behavior and outcomes might emerge from the interactions of learning agents in artificial societies [ de Cote et al, 2006;Foerster et al, 2018;Hughes et al, 2018;Jaques et al, 2019;Leibo et al, 2017;McKee et al, 2020;Peysakhovich and Lerer, 2018a;Peysakhovich and Lerer, 2018b;Rodriguez-Soto et al, 2021;Sandholm and Crites, 1996] and in interactive systems where humans are in the loop [Carroll et al, 2019;Rahwan et al, 2019]. We believe that a promising and insightful starting point is the analysis of emergent behavior of Reinforcement Learning (RL) agents that act according to a predefined set of moral rewards in situations were there is tension between individual interest and collective social outcomes, namely in social dilemmas [Axelrod and Hamilton, 1981;Rapoport, 1974;Sigmund, 2010].…”