2022
DOI: 10.48550/arxiv.2206.05862
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

X-Risk Analysis for AI Research

Abstract: Artificial intelligence (AI) has the potential to greatly improve society, but as with any powerful technology, it comes with heightened risks and responsibilities. Current AI research lacks a systematic discussion of how to manage long-tail risks from AI systems, including speculative long-term risks. Keeping in mind the potential benefits of AI, there is some concern that building ever more intelligent and powerful AI systems could eventually result in systems that are more powerful than us; some say this is… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 5 publications
(5 citation statements)
references
References 22 publications
(29 reference statements)
0
5
0
Order By: Relevance
“…Ethical considerations. An ethical concern one might have about our work is that revealing BNSL might differentially (Hendrycks & Mazeika, 2022) improve A(G)I capabilities progress relative to A(G)I safety/alignment progress. A counter-argument is that BNSL will also allow the A(G)I safety/alignment field to extrapolate the scaling behaviors of its methods for aligning A(G)I systems and as a result will also accelerate alignment/safety progress.…”
Section: Discussionmentioning
confidence: 99%
“…Ethical considerations. An ethical concern one might have about our work is that revealing BNSL might differentially (Hendrycks & Mazeika, 2022) improve A(G)I capabilities progress relative to A(G)I safety/alignment progress. A counter-argument is that BNSL will also allow the A(G)I safety/alignment field to extrapolate the scaling behaviors of its methods for aligning A(G)I systems and as a result will also accelerate alignment/safety progress.…”
Section: Discussionmentioning
confidence: 99%
“…If AI systems are persuasive enough, for instance, if they were to be more persuasive than 99% of humans, their unregulated proliferation could lead to serious degradation in discourse between humans [28]. If systems are persuasive, this could lead to a degradation of truth or potentially reduce trust between humans and machines.…”
Section: Ai-driven Persuasion Could Contribute To a Loss Of Human Con...mentioning
confidence: 99%
“…11 Talking points include problems such as: Weaponization, where governments are strongly incentivized to weaponize AI, which would significantly increase the risks of conflict; 12 Enfeeblement, where important decisions may be handed off to AI, endangering humanity's capacity for selfgovernance. (This scenario was depicted in the film WALL-E); 13 Eroded epistemics, where nations, political parties, and many other actors are strongly incentivized to develop agents that spread propaganda, undermining our ability to seek truth; 14 Proxy gaming, where AI may strongly shape human behavior in suboptimal ways, illustrated by addiction caused by social media recommendation algorithms; 15 and Value lock-in, where advanced AI lock-in the dominance of the nations or companies that develop it, curtailing capacity for social progress. 16 We will discuss which practices should be implemented to ensure the development of beneficial AI.…”
Section: Part 3: the Difficulty Of Governing Ai [25 Min]mentioning
confidence: 99%