Proceedings of the 22nd Conference on Computational Natural Language Learning 2018
DOI: 10.18653/v1/k18-1037
|View full text |Cite
|
Sign up to set email alerts
|

Evolutionary Data Measures: Understanding the Difficulty of Text Classification Tasks

Abstract: Classification tasks are usually analysed and improved through new model architectures or hyperparameter optimisation but the underlying properties of datasets are discovered on an ad-hoc basis as errors occur. However, understanding the properties of the data is crucial in perfecting models. In this paper we analyse exactly which characteristics of a dataset best determine how difficult that dataset is for the task of text classification. We then propose an intuitive measure of difficulty for text classificat… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
7
0
1

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 16 publications
(8 citation statements)
references
References 50 publications
0
7
0
1
Order By: Relevance
“…N is the class probability for class i, and q i = 1 C denotes the uniform probability. Imbalance Absolute Deviation [4] is defined as the sum of aboslute distance between each long-tailed and uniform probability:…”
Section: Motivation and Metrics For Evaluating Data Imbalancementioning
confidence: 99%
“…N is the class probability for class i, and q i = 1 C denotes the uniform probability. Imbalance Absolute Deviation [4] is defined as the sum of aboslute distance between each long-tailed and uniform probability:…”
Section: Motivation and Metrics For Evaluating Data Imbalancementioning
confidence: 99%
“…For instance, "coarse" sentiment classification involves distinguishing negative and positive sentiments in text, and fine-grained sentiment classification involves further distinguishing the positive class into very positive and positive. This problem is challenging because the classes are semantically similar, which makes it difficult for the model to learn the labels (Collins et al, 2018).…”
Section: Fine-grained Classificationmentioning
confidence: 99%
“…This involves distinguishing between some closely confusable pairs of emotions, such as "sad" and "devastated", or "furious" and "annoyed". Fine-grained classification tasks are challenging precisely due to the presence of class interference amongst closely confusable classes (Collins et al, 2018;Zhao et al, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…Understanding the datasets In machine learning and NLP literature, several works studied the "difficulty" of datasets (Blache and Rauzy, 2011;Gupta et al, 2014;Collins et al, 2018;Jain et al, 2020), but they did not consider factoring out the impacts of shortcuts. D'Amour et al ( 2020) framed the shortcut learning issue as an underspecification problem: There is not enough information in training set to distinguish between spurious artifacts and the inductive biases (or rather, the linguistic knowledge).…”
Section: Related Workmentioning
confidence: 99%
“…Metrics for cross-task comparison. Consider reporting the performance on a unified scale of "task-specific informativeness", rather than relying on average model performance metrics (Collins et al, 2018). Designing metrics with grounds in linguistic knowledge is an interesting direction of future work.…”
Section: Broader Impactsmentioning
confidence: 99%