Recent events that revolve around fake news indicate that humans are more susceptible than ever to mental manipulation by powerful technological tools. In the future these tools may become autonomous. One crucial property of autonomous agents is their potential ability to deceive. From this research we hope to understand the potential risks and benefits of deceptive artificial agents. The method we propose to study deceptive agents is by making them interact with agents that detect deception and analyse what emerges from these interactions given multiple setups such as formalisations of scenarios inspired from historical cases of deception.
Deception and Thinking MachinesThere is some related work in the AI literature that focuses on the issue of deception. Bridewell and Isaac define a model of dynamic belief attribution for deception in [Bridewell and Isaac, 2011]. Jones models self-deception using epistemic logic in [Jones, 2015]. Lambert defines a cognitive model of deception based on human-computer interaction in [Lambert, 1987]. Multiple studies have been done by Sakama and Caminada on formalising dishonesty in [Sakama, 2011], [Sakama et al., 2014] such as deductive and abductive dishonesty, lies, bullshit, and deception. In [Sakama and Caminada, 2010] and [Sakama, 2015] they formalise multiple types of deception.Deceptive machines first appeared as concepts in Turing's Imitation Game. Today it is reasonable to imagine machines that exploit human masses to extract rewards, e.g. deceiving humans into voting for an entity that the machines consider to be a necessary requirement for their success in attaining an ulterior goal. Such autonomous systems might emerge from complex social interactions that they will be able to eventually manipulate according to their will. For example, the system might use crowdturfing attacks to generate fake reviews as demonstrated in [Yao et al., 2017]. Such attacks can be used to maximise the profits of entities that can afford them financially or by entities that consider that legal risks are too low and choose to employ them in spite of the established rules.There are three main reasons for modelling artificial agents. Firstly, deception is fundamental to a complete theory of communication and by modelling deceptive agents we might be able to get a better understanding of how deception works. Secondly, intelligent machines might develop reasons to deceive. Understanding their reasoning and abilities can help us identify and prevent them from deceiving us or other artificial agents [Sakama, 2011] [Castelfranchi andTan, 2001]. Thirdly, deception seems to be a necessary step in developing AI that emulates human cognition.
Approach and FoundationsTwo main paradigms seem to be emerging within the AI community: (i) a model-driven paradigm and (ii) a data-driven paradigm. The model-driven paradigm stands for building an AI that reasons using models which contain beliefs and knowledge about the world and about other agents, in order to interpret evidence (data) and to act ac...