2022
DOI: 10.1038/s41562-022-01383-x
|View full text |Cite|
|
Sign up to set email alerts
|

Human-centred mechanism design with Democratic AI

Abstract: Building artificial intelligence (AI) that aligns with human values is an unsolved problem. Here we developed a human-in-the-loop research pipeline called Democratic AI, in which reinforcement learning is used to design a social mechanism that humans prefer by majority. A large group of humans played an online investment game that involved deciding whether to keep a monetary endowment or to share it with others for collective benefit. Shared revenue was returned to players under two different redistribution me… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
16
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
2
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 47 publications
(16 citation statements)
references
References 35 publications
0
16
0
Order By: Relevance
“…Considering other models may change the system dynamics, and perhaps improve the PoA. Moreover, designing an optimal fairness rule may be done similarly as for voting mechanisms in the manner of [8].…”
Section: Discussionmentioning
confidence: 99%
See 1 more Smart Citation
“…Considering other models may change the system dynamics, and perhaps improve the PoA. Moreover, designing an optimal fairness rule may be done similarly as for voting mechanisms in the manner of [8].…”
Section: Discussionmentioning
confidence: 99%
“…maximal, amount of good bought by a buyer, and the sense of the objective function ensure that it will be exactly this quantity. The constraint (8) imposes that the buyers pay at most in average 𝑝 𝐺 𝑏 for the Good. The constraint (9) forces that the buyers pay at most in average 𝑝 𝑅 𝑏 for the Right.…”
Section: A Hyperparametersmentioning
confidence: 99%
“…Traditionally, a classical model to study the effect of inequality on cooperation is the linear public goods game [ 26 , 27 ]. In this game, individuals decide how much of their endowment to contribute to a public good.…”
Section: An Overview Of the Previous Literature On Asymmetric Public ...mentioning
confidence: 99%
“…DRL from human feedback is when systems or machines learn to behave by using some little assistance from humans, which can actually bring out the best from both humans and the machines [80]. Building AI systems that align with human values, i.e., humancentered AI, in which DRL strategies can be used in the design of social mechanisms preferred by humans, is still an open research topic [81]. What is quite remarkable from the above discussion on human proxies is that most of the solutions to these problems are found through the harmonious combination of RL strategies and hierarchical sensory processing systems.…”
Section: Integrating the Human Proxy With Deep Reinforcement Learningmentioning
confidence: 99%