Proceedings of the 25th International Conference on Intelligent User Interfaces 2020
DOI: 10.1145/3377325.3377486
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic word recommendation to obtain diverse crowdsourced paraphrases of user utterances

Abstract: Building task-oriented bots requires mapping a user utterance to an intent with its associated entities to serve the request. Doing so is not easy since it requires large quantities of high-quality and diverse training data to learn how to map all possible variations of utterances with the same intent. Crowdsourcing may be an effective, inexpensive, and scalable technique for collecting such large datasets. However, the diversity of the results su ers from the priming e ect (i.e. workers are more likely to use… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
2
2

Relationship

4
4

Authors

Journals

citations
Cited by 10 publications
(6 citation statements)
references
References 60 publications
0
6
0
Order By: Relevance
“…Evaluation metrics. The pipeline configurations were evaluated using automatic evaluation metrics commonly used in assessing paraphrase quality [37] To capture the relevance of the generated paraphrases to the input utterance, we use two different metrics. This includes the Bi-Lingual Evaluation Understudy (BLEU) [19], a widely adopted metric that measures the similarity between two given sentences.…”
Section: Methodsmentioning
confidence: 99%
“…Evaluation metrics. The pipeline configurations were evaluated using automatic evaluation metrics commonly used in assessing paraphrase quality [37] To capture the relevance of the generated paraphrases to the input utterance, we use two different metrics. This includes the Bi-Lingual Evaluation Understudy (BLEU) [19], a widely adopted metric that measures the similarity between two given sentences.…”
Section: Methodsmentioning
confidence: 99%
“…Prior work has shown that crowdsourced natural language benchmarks exhibit various spurious biases (i.e., unintended correlations between input and output), that lead to overestimation of PLM performance [SSK * 17,PNH * 18,GSL * 18,LBSB * 20]. Several techniques have been proposed to handle such bias post‐creation, including improving linguistic diversity of samples [YZFBC * 20, LML * 19, SYH20] and augmenting data with adversarial samples intended to fool the model [WRF * 19,KBN * 21,TYLB * ]. Similarly, there is evidence that natural language instructions provided by dataset creators during crowdsourcing influences crowdworkers to follow specific patterns during sample creation [GGB19,PMGB22, HSG * 21].…”
Section: Related Workmentioning
confidence: 99%
“…This technique is used by Chklovski (2005) where paraphrases are collected via a game where users must reformulate a given sentence based on hints. To collect more diverse paraphrases, Yaghoub-Zadeh-Fard et al (2020) were inspired by another game called Taboo and gave workers a list of taboo words they were not allowed to include in their paraphrases.…”
Section: Literature Reviewmentioning
confidence: 99%