2020
DOI: 10.48550/arxiv.2005.00816
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

DQI: Measuring Data Quality in NLP

Abstract: Neural language models have achieved human level performance across several NLP datasets. However, recent studies have shown that these models are not truly learning the desired task; rather, their high performance is attributed to overfitting using spurious biases, which suggests that the capabilities of AI systems have been over-estimated. We introduce a generic formula for Data Quality Index (DQI) to help dataset creators create datasets free of such unwanted biases. We evaluate this formula using a recentl… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
2

Relationship

1
1

Authors

Journals

citations
Cited by 2 publications
(2 citation statements)
references
References 49 publications
(104 reference statements)
0
2
0
Order By: Relevance
“…To mitigate the bias, prior works have focused on priming crowdsourcing annotators with minimal information to increase their imagination (Geva et al, 2021; to avoid recurring patterns. Arunkumar et al (2020) develops a real time feedback and metric-in-the loop (Mishra et al, 2020b) workflow to educate crowdworkers in controlling dataset biases. provides an iterative protocol with expert assessments for crowdsourcing data collection to increase difficulty of instances.…”
Section: A Biases In Nlu Benchmarksmentioning
confidence: 99%
“…To mitigate the bias, prior works have focused on priming crowdsourcing annotators with minimal information to increase their imagination (Geva et al, 2021; to avoid recurring patterns. Arunkumar et al (2020) develops a real time feedback and metric-in-the loop (Mishra et al, 2020b) workflow to educate crowdworkers in controlling dataset biases. provides an iterative protocol with expert assessments for crowdsourcing data collection to increase difficulty of instances.…”
Section: A Biases In Nlu Benchmarksmentioning
confidence: 99%
“…Dasgupta et al (2018) incorporate compositional information into sentence embeddings for Natural Language Inference. DQI (Mishra et al, 2020) offers quantitative metrics to assess biases in automated dataset creation in Natural Language Processing. Le Bras et al ( 2020) introduce adversarial measures to mitigate biases in various Natural Language Processing and Computer Vision tasks.…”
Section: Related Workmentioning
confidence: 99%