2023 ACM Conference on Fairness, Accountability, and Transparency 2023
DOI: 10.1145/3593013.3593995
|View full text |Cite
|
Sign up to set email alerts
|

The ethical ambiguity of AI data enrichment: Measuring gaps in research ethics norms and practices

Abstract: The technical progression of artificial intelligence (AI) research has been built on breakthroughs in fields such as computer science, statistics, and mathematics. However, in the past decade AI researchers have increasingly looked to the social sciences, turning to human interactions to solve the challenges of model development. Paying crowdsourcing workers to generate or curate data, or 'data enrichment', has become indispensable for many areas of AI research, from natural language processing to reinforcemen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
2
1
1

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 45 publications
0
4
0
Order By: Relevance
“…RLHF workers can provide subjective perceptions of the truthfulness or accuracy of statements and indicate a preference for factual responses. 4 The ever-popular solution of 'more data' can make LLMs sound more convincing without necessarily increasing their reliability ( [5], pp. 9-11, [48,49]).…”
Section: Truth and Large Language Modelsmentioning
confidence: 99%
See 3 more Smart Citations
“…RLHF workers can provide subjective perceptions of the truthfulness or accuracy of statements and indicate a preference for factual responses. 4 The ever-popular solution of 'more data' can make LLMs sound more convincing without necessarily increasing their reliability ( [5], pp. 9-11, [48,49]).…”
Section: Truth and Large Language Modelsmentioning
confidence: 99%
“…2-6], LLMs would need to transform from incidental to deterministic truth tellers. 4 It is worth noting, however, that RLHF and other data enrichment work are often highly opaque and secretive processes. Users will rarely have full awareness of how precisely models have been aligned with ground truth through human feedback.…”
Section: Truth and Large Language Modelsmentioning
confidence: 99%
See 2 more Smart Citations