2022 ACM Conference on Fairness, Accountability, and Transparency 2022
DOI: 10.1145/3531146.3533229
|View full text |Cite
|
Sign up to set email alerts
|

Predictability and Surprise in Large Generative Models

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

1
79
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

0
9

Authors

Journals

citations
Cited by 113 publications
(80 citation statements)
references
References 10 publications
1
79
0
Order By: Relevance
“…Leaderboards that integrate conventional benchmarking with analysis can be especially helpful by making this largely automatic (Wang et al, 2018;Dua et al, 2019;Gehrmann et al, 2021;. More broadly, careful analysis work, targeted at broadly understanding the capacities of capable models, will be valuable in helping to forecast and mitigate the worst risks from future systems (Elhage et al, 2021;Ganguli et al, 2022).…”
Section: Ways To Do Bettermentioning
confidence: 99%
“…Leaderboards that integrate conventional benchmarking with analysis can be especially helpful by making this largely automatic (Wang et al, 2018;Dua et al, 2019;Gehrmann et al, 2021;. More broadly, careful analysis work, targeted at broadly understanding the capacities of capable models, will be valuable in helping to forecast and mitigate the worst risks from future systems (Elhage et al, 2021;Ganguli et al, 2022).…”
Section: Ways To Do Bettermentioning
confidence: 99%
“…Novel AI capabilities continue to emerge, 76 increasing the urgency to align AI with humans. Legal standards can serve as a pillar of AI goal specification practices.…”
Section: Discussionmentioning
confidence: 99%
“…Value lock-in could occur when our technology perpetuates the values of a particular powerful group, or it could occur when groups get stuck in a poor equilibrium that is robust to attempts to get unstuck. Emergent functionality could be hazardous because models demonstrate unexpected, qualitatively different behavior as they become more competent [26,57], so a loss of control becomes more likely when new capabilities or goals spontaneously emerge. Deception is commonly incentivized, and smarter agents are more capable of succeeding at deception; we can be less sure of our models if we fail to find a way to make them assert only what they hold to be true.…”
Section: Speculative Hazards and Failure Modesmentioning
confidence: 99%
“…6. Emergent functionality: Capabilities and novel functionality can spontaneously emerge in today's AI systems [26,57], even though these capabilities were not anticipated by system designers. If we do not know what capabilities systems possess, systems become harder to control or safely deploy.…”
mentioning
confidence: 99%