Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing 2023
DOI: 10.18653/v1/2023.emnlp-main.117
|View full text |Cite
|
Sign up to set email alerts
|

ChatGPT to Replace Crowdsourcing of Paraphrases for Intent Classification: Higher Diversity and Comparable Model Robustness

Jan Cegin,
Jakub Simko,
Peter Brusilovsky

Abstract: The emergence of generative large language models (LLMs) raises the question: what will be its impact on crowdsourcing? Traditionally, crowdsourcing has been used for acquiring solutions to a wide variety of human-intelligence tasks, including ones involving text generation, modification or evaluation. For some of these tasks, models like ChatGPT can potentially substitute human workers. In this study, we investigate whether this is the case for the task of paraphrase generation for intent classification. We a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
3
2

Relationship

0
5

Authors

Journals

citations
Cited by 6 publications
(1 citation statement)
references
References 20 publications
0
1
0
Order By: Relevance
“…These situations are typical when computer science researchers develop datasets in specialized fields like biomedicine [34,36,39,40,46,67]. 8 In such instances, experts within the domain often are unable to label large datasets quickly, so out-of-domain experts-frequently CS graduate students-sometime need to label portions of the data to assess the crowd labels' quality. Second, the accuracy of the CS Expert's labels sets an estimated upper limit for non-expert performance, given their familiarity with the task and focused attention.…”
Section: Collecting Gold-standard Labels Using Expertsmentioning
confidence: 99%
“…These situations are typical when computer science researchers develop datasets in specialized fields like biomedicine [34,36,39,40,46,67]. 8 In such instances, experts within the domain often are unable to label large datasets quickly, so out-of-domain experts-frequently CS graduate students-sometime need to label portions of the data to assess the crowd labels' quality. Second, the accuracy of the CS Expert's labels sets an estimated upper limit for non-expert performance, given their familiarity with the task and focused attention.…”
Section: Collecting Gold-standard Labels Using Expertsmentioning
confidence: 99%