2020
DOI: 10.2478/jagi-2020-0004
|View full text |Cite
|
Sign up to set email alerts
|

The Archimedean trap: Why traditional reinforcement learning will probably not yield AGI

Abstract: After generalizing the Archimedean property of real numbers in such a way as to make it adaptable to non-numeric structures, we demonstrate that the real numbers cannot be used to accurately measure non-Archimedean structures. We argue that, since an agent with Artificial General Intelligence (AGI) should have no problem engaging in tasks that inherently involve non-Archimedean rewards, and since traditional reinforcement learning rewards are real numbers, therefore traditional reinforcement learning probably … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 8 publications
(6 citation statements)
references
References 23 publications
0
6
0
Order By: Relevance
“…However, to simulate feelings that result in human-like behaviour is a more difficult proposition. Rather than trying to describe human-like feelings, we simplify our analysis by assuming the preferences [21] n o which are determined by experience of feelings.…”
Section: Extending the Formalismmentioning
confidence: 99%
“…However, to simulate feelings that result in human-like behaviour is a more difficult proposition. Rather than trying to describe human-like feelings, we simplify our analysis by assuming the preferences [21] n o which are determined by experience of feelings.…”
Section: Extending the Formalismmentioning
confidence: 99%
“…However, to simulate feelings that result in human-like behaviour is a more difficult proposition. Rather than trying to describe human-like feelings, we simplify our analysis by assuming the preferences [20] n o which are determined by experience of feelings.…”
Section: Extending the Formalismmentioning
confidence: 99%
“…These urges impose an order of preference on decisions, and a threshold beyond which the utility is "good enough" to indicate completion [18,27]. This ordering is akin to preference based reinforcement learning [39], which allows adaptation to "non-Archimedean tasks whose rewards are too rich to express using real numbers" [40]. Given preference, decisions can be evaluated and ostensive definitions constructed.…”
Section: Ambiguity and Correctnessmentioning
confidence: 99%