2022
DOI: 10.48550/arxiv.2201.12323
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Describing Differences between Text Distributions with Natural Language

Abstract: How do two distributions of text differ? Humans are slow at answering this, since discovering patterns might require tediously reading through hundreds of samples. We propose to automatically summarize the differences by "learning a natural language hypothesis": given two distributions D 0 and D 1 , we search for a description that is more often true for D 1 , e.g., "is military-related." To tackle this problem, we fine-tune GPT-3 to propose descriptions with the prompt: "[samples of D 0 ] + [samples of D 1 ] … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 39 publications
(53 reference statements)
0
1
0
Order By: Relevance
“…Scalable Oversight As AI systems become more capable of generating candidate responses, an emerging line of research supervises AI systems by providing preferences over AI-generated candidates rather than providing human demonstrations (Stiennon et al, 2020;Wiegreffe et al, 2021;Askell et al, 2021;Liu et al, 2022;Ouyang et al, 2022). Therefore, to supervise AI to perform more complex tasks, it becomes increasingly important to determine human preferences over model outputs that are expensive to verify, such as full-book summaries or natural language descriptions of distributional properties (Amodei et al, 2016;Wu et al, 2021;Zhong et al, 2022). Our work presents a strategy to re-weight complex outputs (programs) from an AI system by asking simple informative questions of annotators who do not have to understand the outputs directly.…”
Section: Related Workmentioning
confidence: 99%
“…Scalable Oversight As AI systems become more capable of generating candidate responses, an emerging line of research supervises AI systems by providing preferences over AI-generated candidates rather than providing human demonstrations (Stiennon et al, 2020;Wiegreffe et al, 2021;Askell et al, 2021;Liu et al, 2022;Ouyang et al, 2022). Therefore, to supervise AI to perform more complex tasks, it becomes increasingly important to determine human preferences over model outputs that are expensive to verify, such as full-book summaries or natural language descriptions of distributional properties (Amodei et al, 2016;Wu et al, 2021;Zhong et al, 2022). Our work presents a strategy to re-weight complex outputs (programs) from an AI system by asking simple informative questions of annotators who do not have to understand the outputs directly.…”
Section: Related Workmentioning
confidence: 99%