2021
DOI: 10.1093/pq/pqaa086
|View full text |Cite
|
Sign up to set email alerts
|

Extracting Money from Causal Decision Theorists

Abstract: Newcomb’s problem has spawned a debate about which variant of expected utility maximisation (if any) should guide rational choice. In this paper, we provide a new argument against what is probably the most popular variant: causal decision theory (CDT). In particular, we provide two scenarios in which CDT voluntarily loses money. In the first, an agent faces a single choice and following CDT’s recommendation yields a loss of money in expectation. The second scenario extends the first to a diachronic Dutch book … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2023
2023
2024
2024

Publication Types

Select...
4
1

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(4 citation statements)
references
References 26 publications
0
4
0
Order By: Relevance
“…Hence, there exists no agent that complies with standard BDT in this problem. Compare the example of Oesterheld and Conitzer [18] and Spencer [28]; also see Demski and Garrabrant ([7], Sect. 2.1) for a discussion of another, subtler issue that arises from logical omniscience and introspection.…”
Section: Computational Constraints and Paradoxes Of Self-referencementioning
confidence: 99%
See 1 more Smart Citation
“…Hence, there exists no agent that complies with standard BDT in this problem. Compare the example of Oesterheld and Conitzer [18] and Spencer [28]; also see Demski and Garrabrant ([7], Sect. 2.1) for a discussion of another, subtler issue that arises from logical omniscience and introspection.…”
Section: Computational Constraints and Paradoxes Of Self-referencementioning
confidence: 99%
“…Paradoxes of self-reference, strategic interactions, and counterfactuals A second problem with BDT and logical omniscience more generally is that it creates inconsistencies if the values of different available options depend on what the agent chooses. As an example, consider the following decision problem, which we will call the Simplified Adversarial Offer (SAO) (after a decision problem introduced by [18]). Imagine that an artificial agent chooses between two available alternatives a 0 and a 1 , where a 0 is known to pay off 1 /2 with certainty, and a 1 is known to pay off 1 if the agent's program run on this decision problem chooses a 0 , and 0 otherwise.…”
Section: Computational Constraints and Paradoxes Of Self-referencementioning
confidence: 99%
“…For example, Bell et al (2021) show that softmax Q-learners behave more like CDT agents. Albert and Heiner (2001), Mayer, Feldmaier, and Shen (2016), Oesterheld, Demski, and Conitzer (2021), and describe methods of learning that result in EDT-like behavior (see also Oesterheld 2021). We hope that future work in this area will shed light on the feasibility of building learning agents that cooperate against near-copies in real-world, complex, asymmetric scenarios.…”
Section: Cooperation Between Copiesmentioning
confidence: 99%
“…Another unique type of failure mode for CDT can be referred to as the money pump dilemma [48]. This involves an agent voluntarily playing a game against a predictor, Omega, who has access to their source code.…”
Section: Predictormentioning
confidence: 99%